WO2020248386A1 - Video analysis method and apparatus, computer device and storage medium - Google Patents

Video analysis method and apparatus, computer device and storage medium Download PDF

Info

Publication number
WO2020248386A1
WO2020248386A1 PCT/CN2019/103373 CN2019103373W WO2020248386A1 WO 2020248386 A1 WO2020248386 A1 WO 2020248386A1 CN 2019103373 W CN2019103373 W CN 2019103373W WO 2020248386 A1 WO2020248386 A1 WO 2020248386A1
Authority
WO
WIPO (PCT)
Prior art keywords
target object
abnormal
image
video
video image
Prior art date
Application number
PCT/CN2019/103373
Other languages
French (fr)
Chinese (zh)
Inventor
盖超
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2020248386A1 publication Critical patent/WO2020248386A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7837Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream

Definitions

  • This application relates to the field of image recognition technology, and in particular to a video analysis method, device, computer equipment and storage medium.
  • the first aspect of the present application provides a video analysis method, the method includes:
  • the detecting the target object in the video image to obtain the target object category includes:
  • the tracking the target object in the video image to obtain the state of the target object includes:
  • the target object does not appear in the detection range in the current video frame, it is determined that the target object is abnormal.
  • the judging whether the business scenario is abnormal includes:
  • the abnormal model When the abnormal model outputs the abnormal scene corresponding to the image to be recognized, it is confirmed that the business scene is abnormal.
  • the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
  • the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
  • the method further includes:
  • the third-party business platform includes a public security system and a traffic control system.
  • the method further includes:
  • a second aspect of the present application provides a video analysis device, the device includes:
  • the receiving module is used to receive the video image collected by the camera
  • the detection module is used to detect the target object in the video image to obtain the target object category
  • a tracking module used to track the target object in the video image to obtain the state of the target object
  • An analysis module configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the state of the target object;
  • the judgment module is used to judge whether the business scenario is abnormal.
  • the processing module is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
  • a third aspect of the present application provides a computer device that includes a processor and a memory, and the processor is configured to implement the video analysis method when executing computer-readable instructions stored in the memory.
  • a fourth aspect of the present application provides a non-volatile readable storage medium having computer readable instructions stored on the non-volatile readable storage medium, and when the computer readable instructions are executed by a processor, the Video analysis methods.
  • the video analysis method, device, computer equipment and storage medium described in this application can analyze the video image to obtain the business scenario contained in the video image, and determine whether the business scenario is abnormal, and when the business scenario is abnormal , To record the key information when the business scenario is abnormal. Thereby, the key information can be sent to the corresponding third-party platform, and the exception can be handled in time.
  • FIG. 1 is a flowchart of a video analysis method provided in Embodiment 1 of the present application.
  • FIG. 2 is a diagram of functional modules in a preferred embodiment of the video analysis device of this application provided in the second embodiment of this application.
  • Fig. 3 is a schematic diagram of a computer device provided in Embodiment 3 of the present application.
  • the video analysis method of the embodiment of the present application is applied in a hardware environment composed of at least one computer device and a mobile terminal connected to the computer device through a network.
  • Networks include but are not limited to: wide area network, metropolitan area network or local area network.
  • the video analysis method in the embodiments of the present application may be executed by a computer device or a mobile terminal; it may also be executed by the computer device and the mobile terminal.
  • the video analysis function provided by the method of this application can be directly integrated on the computer device, or a client for implementing the method of this application can be installed.
  • the method provided in this application can also be run on a computer or other device in the form of a software development kit (SDK), and provide an interface for video analysis functions in the form of SDK. The interface can realize the video analysis function.
  • SDK software development kit
  • FIG. 1 is a flowchart of a video analysis method provided in Embodiment 1 of the present application. According to different needs, the execution order in this flowchart can be changed, and some steps can be omitted.
  • Step S1 receiving the video image collected by the camera.
  • the video image is collected by a camera, and the camera is installed in different business scenarios.
  • the business scenario describes a scenario that requires target object detection and/or video analysis.
  • the business scenario is an intelligent traffic business scenario that recognizes traffic accidents, congestion, vehicle speed detection, traffic flow prediction, vehicle loss of control, vehicle trajectory, people or bicycle intrusion, violation of traffic laws, throwing objects, etc.
  • the business scenario also It can be a smart park business scenario that identifies personnel intrusion, leftovers, lost property monitoring, license plate analysis, vehicle trajectory, traffic flow analysis, pedestrian flow analysis, fireworks or smoke, etc.
  • the business scenario can also be illegal ships, overloaded, dense crowds, etc. Ferry monitoring business scenarios such as detection, whether to wear a life jacket, falling into the water, etc.
  • the business scenarios may also be scenarios such as unmanned driving, financial scenarios, equipment login, airport and public area monitoring.
  • the cameras may be cameras of different models and specifications manufactured by different manufacturers, and the video analysis method can realize unified processing and analysis of video images taken by cameras of different models and specifications manufactured by different manufacturers.
  • the video analysis method further includes:
  • the video image may be video decoded by a graphics processing unit (GPU) to obtain each frame of the video image.
  • GPU graphics processing unit
  • Step S2 Detect the target object in the video image to obtain the target object category.
  • the target objects in the video image include people, animals, vehicles, buildings, smoke and so on.
  • detecting the target object in the video image to obtain the target object category includes:
  • the target object in the video image includes a static target object and a moving target object.
  • the stationary target object can be identified through a template-based detection method. Specifically, it includes: determining the contour of the target object shape in the video image, and matching the contour of the target object shape with a pre-stored template file.
  • the outline of the shape of the target object can be determined to be a rectangle, and the rectangle is feature-matched with a pre-stored door template file to identify the target object.
  • the template file of the door is rectangular.
  • the target object in the video image is a moving target object
  • it can be identified by at least one of the background difference method, the frame difference method, and the optical flow method.
  • the background difference method is to perform background modeling on a relatively fixed scene in the video image, and the moving target object is obtained from the difference between the current image and the background model during detection; The corresponding position pixels between the frames are compared to obtain the position of the moving target object; the optical flow method uses the time-varying optical flow vector characteristics to detect the moving target object in the video image.
  • the above-mentioned method for detecting static target objects and moving target objects in a video image is not limited to the above-mentioned enumeration, and any reproduction method suitable for detecting a target object in a video image can be applied to this.
  • the methods for detecting stationary target objects and moving target objects in a video image in this embodiment are all existing technologies, and will not be described in detail herein.
  • the target object in the video image when the target object in the video image is recognized as a car, it may be determined that the target object is a vehicle.
  • the detection and classification of the target object is a very basic task in vision technology, and its purpose is to track some objects of interest in the scene, including conventional target object detection, person detection, vehicle detection, and so on.
  • the basic attributes of the target object in the video image can be obtained by decomposing the target object in the video image, where the basic attributes include color, motion track, shape, structure, etc., and then The obtained basic attributes are compared with the basic attributes of the target object pre-stored in the database, so as to accurately identify the target object in the video image.
  • the database stores a table corresponding to the basic attributes of the target object and the target object category.
  • the determining the category of the target object specifically includes: obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image; and comparing the obtained basic attributes with the ones stored in a database in advance.
  • the basic attributes of the target object are compared; when the acquired basic attributes are consistent with the basic attributes of the target object in the database, the corresponding table of basic attributes and target object categories stored in the database is queried to obtain the target object’s category.
  • Step S3 tracking the target object in the video image to obtain the state of the target object.
  • the state of the target object can be determined by tracking the target object in the video image.
  • the method for tracking the target object in the video image includes:
  • e) Determine whether the target object appears in the detection range of the current video frame, if the target object does not appear in the detection range of the current video frame, determine that the state of the target object is abnormal; if the target object appears in the detection range
  • the detection range in the current video frame determines the image area of the target object in the current video frame, that is, the state of the target object is normal.
  • the pre-order video frame refers to the k video frames before the current video frame
  • the current video frame is estimated and compared and detected through the first k video frames, which requires a small amount of calculation and can solve the occasional loss of target objects in the video Or the problem of occlusion, the detection accuracy is higher.
  • Step S4 Analyze the business scene contained in the video image according to the category of the target object and the state of the target object.
  • the category of the target object can be obtained according to the detection result, and the state of the target object can be determined according to the tracking result, so that the business scene contained in the video image can be analyzed.
  • the category of the target object is a car. If the car does not appear in the detection range of the current video frame, it can be determined that the state of the car is abnormal. If the car is in a congested state, it can be It is learned that the business scene included in the video image is an intelligent transportation business scene.
  • the category of the target object can be obtained as a pedestrian, and the pedestrian does not appear in the detection range in the current video frame, it can be determined that the state of the pedestrian is abnormal. If the pedestrian falls down, it can be known that the business scene included in the video image is an intelligent traffic business scene.
  • the category of the target object can be obtained as a door according to the detection result, and the door does not appear in the detection range in the current video, it can be confirmed that the status of the door is abnormal. If it is kept open, it can be judged that the door is in the video image.
  • the business scenarios included are smart security business scenarios.
  • Step S5 Determine whether the business scene in the video image is abnormal.
  • step S6 is entered; when the business scene in the video image is not abnormal, the process ends.
  • the video image may be input to a pre-trained abnormality model, and whether the business scene in the video image is abnormal or not can be determined according to the abnormality model. Specifically, when it is determined that the target object is abnormal, extract the current video frame as an abnormal image; import the abnormal image as an image to be recognized into a pre-trained anomaly model, wherein the abnormal model is used to characterize the Identify the correspondence between the image and the abnormal scene; when the abnormal model outputs the abnormal scene corresponding to the image to be identified, confirm that the business scene is abnormal and the abnormal model includes abnormal models corresponding to different business scenarios.
  • the abnormal model corresponding to the intelligent transportation business scene includes a traffic accident model, a traffic congestion model, and an illegal scale type, etc.
  • the business scene is a smart park business scene
  • the abnormal model corresponding to the business scene of the smart park includes a personal belongings model, a personnel intrusion model, etc.
  • the abnormal model corresponding to the ferry monitoring business scene includes an overload model, a falling water model, Illegal ship model, etc.
  • the current video frame is extracted as an abnormal image, and the abnormal image is imported into a pre-trained traffic congestion model as an image to be recognized.
  • the traffic jam model outputs the traffic jam scene corresponding to the image to be recognized, it is confirmed that there is an abnormality in the intelligent transportation business scene corresponding to the video image; when the traffic jam model does not output the traffic corresponding to the image to be recognized In a congested scene, confirm that the intelligent transportation service scene corresponding to the video image is normal.
  • the above-mentioned abnormal model is a machine learning model trained based on a picture sample set.
  • the picture samples include abnormal business scene picture samples and normal business scene picture samples.
  • the machine learning model is an artificial intelligence algorithm model that can perform image recognition, including: a convolutional neural network model CNN, a recurrent neural network module RNN, and a deep neural network model DNN.
  • the convolutional neural network model CNN is a multi-layer neural network, which can continuously reduce the dimensionality of the image recognition problem with a huge amount of data, and finally enable it to be trained. Therefore, the machine learning model in the embodiment of the present application may be CNN model.
  • the ResNet network proposes a residual learning framework that reduces the burden of network training. This network is inherently deeper than the previously used network, and solves the problem of other neural networks that decrease in accuracy as the network deepens.
  • the machine learning model may be the ResNet model in the convolution application network model CNN. It should be noted that this is only an example, and other machine learning models that can perform image recognition are also applicable to this application, and will not be repeated here.
  • Step S6 When the business scene in the video image is abnormal, record the key information when the business scene is abnormal.
  • the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
  • the video analysis method further includes sending the recorded key information to a third-party service platform.
  • the third-party service platform includes a public security system, a traffic control system, etc.
  • the third-party business platform can help the third-party business platform to obtain key information when an exception occurs in a business scenario in time, so as to process the exception in time.
  • the video analysis method further includes: displaying key information when the business scenario is abnormal. Specifically, information such as an abnormal picture, time, and point of the business scene is displayed on the display screen.
  • the video analysis method includes receiving a video image collected by a camera; detecting a target object in the video image to obtain the category of the target object; tracking the target object in the video image to obtain the The status of the target object; the business scene contained in the video image is obtained by analyzing the category of the target object and the status of the target object; judging whether the business scene is abnormal; and when the business scene in the video image
  • record the key information when the business scenario is abnormal It is possible to analyze in real time whether the business scene corresponding to the video image is abnormal, and when it is confirmed that the business scene is abnormal, record the key information when the business scene is abnormal. Thereby, the key information can be sent to the corresponding third-party platform, and the exception can be handled in time.
  • FIG. 2 is a diagram of functional modules in a preferred embodiment of the video analysis device of this application.
  • the video analysis device 20 runs in a computer device.
  • the video analysis device 20 may include multiple functional modules composed of computer-readable instruction code segments.
  • the instruction codes of each computer-readable instruction code segment in the video analysis device 20 can be stored in a memory and executed by at least one processor to perform (see FIG. 1 and related descriptions for details) video analysis functions.
  • the video analysis device 20 can be divided into multiple functional modules according to the functions it performs.
  • the functional modules may include: a receiving module 201, a detection module 202, a tracking module 203, an analysis module 204, a judgment module 205, and a processing module 206.
  • the module referred to in this application refers to a series of computer-readable instruction code segments that can be executed by at least one processor and can complete fixed functions, and are stored in a memory. In some embodiments, the functions of each module will be detailed in subsequent embodiments.
  • the receiving module 201 is used to receive video images collected by a camera.
  • the video image is collected by a camera, and the camera is installed in different business scenarios.
  • the business scenario describes a scenario that requires target object detection and/or video analysis.
  • the business scenario is an intelligent traffic business scenario that recognizes traffic accidents, congestion, vehicle speed detection, traffic flow prediction, vehicle loss of control, vehicle trajectory, people or bicycle intrusion, violation of traffic laws, throwing objects, etc.
  • the business scenario also It can be a smart park business scenario that identifies personnel intrusion, leftovers, lost property monitoring, license plate analysis, vehicle trajectory, traffic flow analysis, pedestrian flow analysis, fireworks or smoke, etc.
  • the business scenario can also be illegal ships, overloaded, dense crowds, etc. Ferry monitoring business scenarios such as detection, whether to wear a life jacket, falling into the water, etc.
  • the camera and the computer device are connected through a wired or wireless network communication.
  • the camera sends the collected video images to the computer device through a wired or wireless network.
  • the business scenarios may also be scenarios such as unmanned driving, financial scenarios, equipment login, airport and public area monitoring.
  • the cameras may be cameras of different models and specifications manufactured by different manufacturers
  • the video analysis device 20 may implement unified processing and analysis of video images taken by cameras of different models and specifications manufactured by different manufacturers.
  • the video analysis device 20 may also decode the video image.
  • the video image may be video decoded by a graphics processing unit (GPU) to obtain each frame of the video image.
  • GPU graphics processing unit
  • the detection module 202 is used to detect the target object in the video image to obtain the target object category.
  • the target objects in the video image include people, animals, vehicles, buildings, smoke and so on.
  • detecting the target object in the video image to obtain the target object category includes:
  • the target object in the video image includes a static target object and a moving target object.
  • the stationary target object can be identified through a template-based detection method. Specifically, it includes: determining the contour of the target object shape in the video image, and performing feature matching between the contour of the target object shape and a pre-stored template file.
  • the target object in the video image is a moving target object
  • it can be identified by at least one of the background difference method, the frame difference method, and the optical flow method.
  • the background difference method is to perform background modeling on a relatively fixed scene in the video image, and the moving target object is obtained from the difference between the current image and the background model during detection;
  • the frame difference method is to obtain the moving target object from the difference between the current image and the background model;
  • the corresponding position pixels between the frames are compared to obtain the position of the moving target object;
  • the optical flow method uses the time-varying optical flow vector characteristics to detect the moving target object in the video image.
  • the method for detecting the stationary target object and the moving target object in the video image is not limited to the above-listed method, and any reproduction method suitable for detecting the target object in the video image can be applied to this.
  • the methods for detecting stationary target objects and moving target objects in a video image in this embodiment are all existing technologies, and will not be described in detail herein.
  • the target object in the video image when the target object in the video image is recognized as a car, it may be determined that the target object is a vehicle.
  • the detection and classification of the target object is a very basic task in vision technology, and its purpose is to track some objects of interest in the scene, including conventional target object detection, person detection, vehicle detection, and so on.
  • the basic attributes of the target object in the video image can be obtained by decomposing the target object in the video image, where the basic attributes include color, motion track, shape, structure, etc., and then The obtained basic attributes are compared with the basic attributes of the target object pre-stored in the database, so as to accurately identify the target object in the video image.
  • the database stores a table corresponding to the basic attributes of the target object and the target object category.
  • the determining the category of the target object specifically includes: obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image; and comparing the obtained basic attributes with the ones stored in a database in advance.
  • the basic attributes of the target object are compared; when the acquired basic attributes are consistent with the basic attributes of the target object in the database, the corresponding table of basic attributes and target object categories stored in the database is queried to obtain the target object’s category.
  • the tracking module 203 is configured to track the target object in the video image to obtain the state of the target object.
  • the state of the target object can be determined by tracking the target object in the video image.
  • the method for tracking the target object in the video image includes:
  • e) Determine whether the target object appears in the detection range of the current video frame, if the target object does not appear in the detection range of the current video frame, determine that the state of the target object is abnormal; if the target object appears in the detection range
  • the detection range in the current video frame determines the image area of the target object in the current video frame.
  • the pre-order video frame refers to the k video frames before the current video frame
  • the current video frame is estimated and compared and detected through the first k video frames, which requires a small amount of calculation and can solve the occasional loss of target objects in the video Or the problem of occlusion, the detection accuracy is higher.
  • the analysis module 204 is configured to analyze and obtain the business scenario contained in the video image according to the category of the target object and the state of the target object.
  • the category of the target object can be obtained according to the detection result, and the state of the target object can be determined according to the tracking result, so that the business scene contained in the video image can be analyzed.
  • the target object is a car. If the car does not appear in the detection range of the current video frame, it can be determined that the state of the car is abnormal. If the car is in a congested state, it can be known
  • the business scene included in the video image is an intelligent transportation business scene.
  • the category of the target object can be obtained as a pedestrian, and the pedestrian does not appear in the detection range in the current video frame, it can be determined that the state of the pedestrian is abnormal. If the pedestrian falls down, it can be known that the business scene included in the video image is an intelligent traffic business scene.
  • the judgment module 205 is used to judge whether the business scene in the video image is abnormal. When the business scene in the video image is abnormal, the key information when the business scene is abnormal is recorded.
  • the video image may be input to a pre-trained abnormality model, and whether the business scene in the video image is abnormal or not can be determined according to the abnormality model. Specifically, when it is determined that the target object is abnormal, extract the current video frame as an abnormal image; import the abnormal image as an image to be recognized into a pre-trained abnormal model, where the abnormal model is used to characterize the Identify the correspondence between the image and the abnormal scene; when the abnormal model outputs the abnormal scene corresponding to the image to be identified, confirm that the business scene is abnormal and the abnormal model includes abnormal models corresponding to different business scenarios.
  • the abnormal model corresponding to the intelligent transportation business scenario includes a traffic accident model, a traffic congestion model, and an illegal scale type, etc.
  • the business scenario is a smart park business scenario
  • the abnormal model corresponding to the business scene of the smart park includes a personal belongings model, a personnel intrusion model, etc.
  • the abnormal model corresponding to the ferry monitoring business scene includes an overload model, a falling water model, Illegal ship model, etc.
  • the current video frame is extracted as an abnormal image, and the abnormal image is imported into a pre-trained traffic congestion model as an image to be recognized.
  • the traffic jam model outputs the traffic jam scene corresponding to the image to be recognized, it is confirmed that there is an abnormality in the intelligent transportation business scene corresponding to the video image; when the traffic jam model does not output the traffic corresponding to the image to be recognized In a congested scene, confirm that the intelligent transportation service scene corresponding to the video image is normal.
  • the above-mentioned abnormal model is a machine learning model trained based on a picture sample set.
  • the picture samples include abnormal business scene picture samples and normal business scene picture samples.
  • the machine learning model is an artificial intelligence algorithm model that can perform image recognition, including: a convolutional neural network model CNN, a recurrent neural network module RNN, and a deep neural network model DNN.
  • the convolutional neural network model CNN is a multi-layer neural network, which can continuously reduce the dimensionality of the image recognition problem with a huge amount of data, and finally enable it to be trained. Therefore, the machine learning model in the embodiment of the present application may be CNN model.
  • the ResNet network proposes a residual learning framework that reduces the burden of network training. This network is inherently deeper than the previously used network, and solves the problem of other neural networks that decrease in accuracy as the network deepens.
  • the machine learning model may be the ResNet model in the convolution application network model CNN. It should be noted that this is only an example, and other machine learning models that can perform image recognition are also applicable to this application, and will not be repeated here.
  • the processing module 206 is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
  • the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
  • the video analysis device 20 can also send the recorded key information to a third-party service platform.
  • the third-party service platform includes a public security system, a traffic control system, etc.
  • the third-party business platform can help the third-party business platform to obtain key information when an exception occurs in a business scenario in time, so as to process the exception in time.
  • the video analysis device 20 can also display key information when the business scene is abnormal. Specifically, the display screen displays the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
  • the video analysis device 20 includes a receiving module 201, a detection module 202, a tracking module 203, an analysis module 204, a judgment module 205, and a processing module 206.
  • the receiving module 201 is used to receive a video image collected by a camera;
  • the detection module 202 is used to detect a target object in the video image to obtain the target object category;
  • the tracking module 203 is used to track the video image
  • the target object in obtains the status of the target object;
  • the analysis module 204 is configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the status of the target object;
  • the judgment module 205 It is used to determine whether the business scene is abnormal; and the processing module 206 is used to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
  • the aforementioned integrated unit implemented in the form of a software function module may be stored in a non-volatile readable storage medium.
  • the above-mentioned software function module is stored in a storage medium, and includes several instructions to make a computer device (which can be a personal computer, a dual-screen device, or a network device, etc.) or a processor to execute the various embodiments of this application Method part.
  • FIG. 3 is a schematic diagram of a computer device provided in Embodiment 3 of this application.
  • the computer device 3 includes: a database 31, a memory 32, at least one processor 33, computer readable instructions 34 stored in the memory 32 and executable on the at least one processor 33, and at least one communication bus 35 .
  • the computer-readable instructions 34 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 32 and executed by the at least one processor 33 Execute to complete this application.
  • the one or more modules/units may be a series of computer-readable instruction segments capable of completing specific functions, and the instruction segments are used to describe the execution process of the computer-readable instructions 34 in the computer device 3.
  • the computer device 3 is a device that can automatically perform numerical calculation and/or information processing according to pre-set or stored instructions. Its hardware includes, but is not limited to, a microprocessor, an application specific integrated circuit (application license Specific Integrated Circuit). , ASIC), programmable gate array (Field-Programmable Gate Array, FPGA), digital processor (Digital Signal Processor, DSP), embedded equipment, etc.
  • ASIC application specific integrated circuit
  • FPGA Field-Programmable Gate Array
  • DSP Digital Signal Processor
  • embedded equipment etc.
  • the schematic diagram 3 is only an example of the computer device 3, and does not constitute a limitation on the computer device 3. It may include more or less components than those shown in the figure, or combine certain components, or different components.
  • the computer device 3 may also include input and output devices, network access devices, buses, etc.
  • the database (Database) 31 is a warehouse built on the computer device 3 to organize, store and manage data according to a data structure. Databases are usually divided into three types: hierarchical database, network database and relational database. In this embodiment, the database 31 is used to store the video images and the like.
  • the at least one processor 33 may be a central processing unit (Central Processing Unit, CPU), or may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSPs), application specific integrated circuits (ASICs). ), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • the processor 33 can be a microprocessor or the processor 33 can also be any conventional processor, etc.
  • the processor 33 is the control center of the computer device 3, and connects the entire computer device 3 with various interfaces and lines. Parts.
  • the memory 32 can be used to store the computer-readable instructions 34 and/or modules/units, and the processor 33 runs or executes the computer-readable instructions and/or modules/units stored in the memory 32, and The data stored in the memory 32 is called to realize various functions of the computer device 3.
  • the memory 32 may mainly include a program storage area and a data storage area.
  • the program storage area may store an operating system, an application program required by at least one function (such as a sound playback function, an image playback function, etc.), etc.;
  • the data (such as audio data, etc.) created according to the use of the computer device 3 and the like are stored.
  • the memory 32 may also include non-volatile memory, such as a hard disk, a memory, a plug-in hard disk, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash Card). , At least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device.
  • non-volatile memory such as a hard disk, a memory, a plug-in hard disk, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash Card).
  • Computer readable instruction codes are stored in the memory 32, and the at least one processor 33 can call the computer readable instruction codes stored in the memory 32 to perform related functions.
  • the various modules (receiving module 201, detection module 202, tracking module 203, analysis module 204, judgment module 205, and processing module 206) described in FIG. 2 are computer-readable instruction codes stored in the memory 32, It is executed by the at least one processor 33 to realize the functions of the various modules to achieve the purpose of video analysis.
  • the receiving module 201 is used to receive video images collected by a camera
  • the detection module 202 is configured to detect a target object in the video image to obtain the target object category;
  • the tracking module 203 is configured to track the target object in the video image to obtain the state of the target object;
  • the analysis module 204 is configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the state of the target object;
  • the judgment module 205 is used to judge whether the business scenario is abnormal.
  • the processing module 206 is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
  • the integrated module/unit of the computer device 3 is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a non-volatile readable storage medium.
  • this application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
  • the computer program can be stored in a non-volatile readable storage medium.
  • the computer program includes computer readable instruction code
  • the computer readable instruction code may be in the form of source code, object code, executable file, or some intermediate form.
  • the non-volatile readable medium may include: any entity or device capable of carrying the computer readable instruction code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) etc.
  • the computer device 3 may also include a power source (such as a battery) for supplying power to various components.
  • the power source may be logically connected to the at least one processor 33 through a power management system, so as to be implemented through a power management system. Manage functions such as charging, discharging, and power management.
  • the power supply may also include one or more DC or AC power supplies, recharging systems, power failure detection circuits, power converters or inverters, power supply status indicators and other arbitrary components.
  • the computer device 3 may also include a Bluetooth module, a Wi-Fi module, etc., which will not be repeated here.
  • the functional units in the various embodiments of the present application may be integrated in the same processing unit, or each unit may exist alone physically, or two or more units may be integrated in the same unit.
  • the above-mentioned integrated unit can be implemented in the form of hardware or in the form of hardware plus software functional modules.

Abstract

A video analysis method, comprising: receiving a video image collected by a camera; detecting a target object in the video image to obtain the type of target object; tracking the target object in the video image to obtain a state of the target object; performing, according to the type of target object and the state of the target object, analysis to obtain a business scene included in the video image; determining whether the business scene is abnormal; and when the business scene in the video image is abnormal, recording key information when the business scene is abnormal. Further provided in the present application are a video analysis apparatus, a computer device and a storage medium. By means of the present application, the key information in the video image when an abnormal event occurs can be acquired, and the abnormal event can be timely processed.

Description

视频分析方法、装置、计算机设备及存储介质Video analysis method, device, computer equipment and storage medium
本申请要求于2019年06月14日提交中国专利局,申请号为201910517477.X发明名称为“视频分析方法、装置、服务器及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on June 14, 2019. The application number is 201910517477.X. The invention title is "Video Analysis Method, Device, Server, and Storage Medium". The entire content is incorporated by reference. In this application.
技术领域Technical field
本申请涉及图像识别技术领域,具体涉及一种视频分析方法、装置、计算机设备及存储介质。This application relates to the field of image recognition technology, and in particular to a video analysis method, device, computer equipment and storage medium.
背景技术Background technique
随着视频监控技术的不断发展,我国目前视频监控在智慧城市、数字城市、智慧园区、智能交通、渡口监测等各类项目得以广泛应用。物联网是智慧城市的基础,视频监控将是核心。然而,在对所述监控视频进行分析时,需要用户重新播放监控视频,一帧一帧查看视频图像,以查找视频图像中的异常事件。需要花费大量时间与人力。With the continuous development of video surveillance technology, my country's current video surveillance is widely used in various projects such as smart cities, digital cities, smart parks, smart transportation, and ferry monitoring. The Internet of Things is the foundation of a smart city, and video surveillance will be the core. However, when analyzing the surveillance video, the user is required to replay the surveillance video and view the video image frame by frame to find abnormal events in the video image. It takes a lot of time and manpower.
发明内容Summary of the invention
鉴于以上内容,有必要提出一种视频分析方法、装置、计算机设备及存储介质,能够及时获取视频图像中出现异常事件时的关键信息。In view of the above, it is necessary to propose a video analysis method, device, computer equipment, and storage medium that can obtain key information when an abnormal event occurs in a video image in time.
本申请的第一方面提供一种视频分析方法,所述方法包括:The first aspect of the present application provides a video analysis method, the method includes:
接收摄像头采集的视频图像;Receive video images collected by the camera;
检测所述视频图像中的目标对象得到所述目标对象的类别;Detecting the target object in the video image to obtain the target object category;
跟踪所述视频图像中的目标对象得到所述目标对象的状态;Tracking the target object in the video image to obtain the state of the target object;
根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;Analyzing the business scene contained in the video image according to the category of the target object and the state of the target object;
判断所述业务场景是否出现异常;及Determine whether the business scenario is abnormal; and
当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。When the business scene in the video image is abnormal, the key information when the business scene is abnormal is recorded.
优选地,所述检测所述视频图像中的目标对象得到所述目标对象的类别包括:Preferably, the detecting the target object in the video image to obtain the target object category includes:
通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性;Obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image;
将获取的所述基本属性与预先存储在数据库中的目标对象的基本属性进行比对;Comparing the acquired basic attributes with the basic attributes of the target object pre-stored in the database;
当获取的所述基本属性与所述数据库中的目标对象的基本属性一致时,查询数据库中存储的基本属性与目标对象类别对应表以得到所述目标对象的类别。When the acquired basic attribute is consistent with the basic attribute of the target object in the database, query a table corresponding to the basic attribute and target object category stored in the database to obtain the target object category.
优选地,所述跟踪所述视频图像中的目标对象得到所述目标对象的状态 包括:Preferably, the tracking the target object in the video image to obtain the state of the target object includes:
确定当前视频帧中的目标对象;Determine the target object in the current video frame;
获取目标对象在前序视频帧中的图像区域以及所述图像区域的图像特征,其中,所述前序视频帧为当前视频帧之前的k个视频帧,k为正整数;Acquiring the image area of the target object in the previous video frame and the image characteristics of the image area, where the previous video frame is k video frames before the current video frame, and k is a positive integer;
根据所述目标对象在前序视频帧中的图像区域,对所述目标对象进行运动估计,确定所述目标对象在当前视频帧的预测区域;Perform motion estimation on the target object according to the image area of the target object in the previous video frame, and determine the prediction area of the target object in the current video frame;
根据所述预测区域确定目标对象在当前视频帧中的检测范围;Determine the detection range of the target object in the current video frame according to the prediction area;
判断所述目标对象是否出现在当前视频帧中的检测范围;Judging whether the target object appears in the detection range of the current video frame;
若所述目标对象出现在当前视频帧中的检测范围,确定所述目标对象在当前视频帧中的图像区域;If the target object appears in the detection range of the current video frame, determine the image area of the target object in the current video frame;
若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象异常。If the target object does not appear in the detection range in the current video frame, it is determined that the target object is abnormal.
优选地,所述判断所述业务场景是否出现异常包括:Preferably, the judging whether the business scenario is abnormal includes:
当确定所述目标对象异常时,提取所述当前视频帧作为异常图像;When it is determined that the target object is abnormal, extract the current video frame as an abnormal image;
将所述异常图像作为待识别图像导入预先训练好的异常模型中,其中,所述异常模型用于表征待识别图像与异常场景之间的对应关系;Importing the abnormal image as an image to be identified into a pre-trained anomaly model, where the abnormal model is used to characterize the correspondence between the image to be identified and the abnormal scene;
当所述异常模型输出与所述待识别图像对应的异常场景时,确认所述业务场景出现异常。When the abnormal model outputs the abnormal scene corresponding to the image to be recognized, it is confirmed that the business scene is abnormal.
优选地,所述关键信息包括所述业务场景出现异常的时间、地点、及截取的所述视频图像中所述业务场景出现异常时的图片文件。Preferably, the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
优选地,所述关键信息包括所述业务场景出现异常的时间、地点、及截取的所述视频图像中所述业务场景出现异常时的图片文件。Preferably, the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
优选地,所述方法还包括:Preferably, the method further includes:
发送记录的关键信息至第三方业务平台,其中,所述第三方业务平台包括公安系统和交通管制系统。Send the recorded key information to a third-party business platform, where the third-party business platform includes a public security system and a traffic control system.
优选地,在接收摄像头采集的视频图像后,所述方法还包括:Preferably, after receiving the video image collected by the camera, the method further includes:
对所述视频图像进行解码。Decoding the video image.
本申请的第二方面提供一种视频分析装置,所述装置包括:A second aspect of the present application provides a video analysis device, the device includes:
接收模块,用于接收摄像头采集的视频图像;The receiving module is used to receive the video image collected by the camera;
检测模块,用于检测所述视频图像中的目标对象得到所述目标对象的类别;The detection module is used to detect the target object in the video image to obtain the target object category;
跟踪模块,用于跟踪所述视频图像中的目标对象得到所述目标对象的状态;A tracking module, used to track the target object in the video image to obtain the state of the target object;
分析模块,用于根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;An analysis module, configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the state of the target object;
判断模块,用于判断所述业务场景是否出现异常;及The judgment module is used to judge whether the business scenario is abnormal; and
处理模块,用于当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。The processing module is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
本申请的第三方面提供一种计算机设备,所述计算机设备包括处理器和 存储器,所述处理器用于执行所述存储器中存储的计算机可读指令时实现所述视频分析方法。A third aspect of the present application provides a computer device that includes a processor and a memory, and the processor is configured to implement the video analysis method when executing computer-readable instructions stored in the memory.
本申请的第四方面提供一种非易失性可读存储介质,所述非易失性可读存储介质上存储有计算机可读指令,所述计算机可读指令被处理器执行时实现所述视频分析方法。A fourth aspect of the present application provides a non-volatile readable storage medium having computer readable instructions stored on the non-volatile readable storage medium, and when the computer readable instructions are executed by a processor, the Video analysis methods.
本申请所述的视频分析方法、装置、计算机设备及存储介质,能够分析视频图像得到所述视频图像所包含的业务场景,并判断所述业务场景是否出现异常,当所述业务场景出现异常时,记录所述业务场景出现异常时的关键信息。从而可以将所述关键信息发送至对应的第三方平台,以及时处理所述异常。The video analysis method, device, computer equipment and storage medium described in this application can analyze the video image to obtain the business scenario contained in the video image, and determine whether the business scenario is abnormal, and when the business scenario is abnormal , To record the key information when the business scenario is abnormal. Thereby, the key information can be sent to the corresponding third-party platform, and the exception can be handled in time.
附图说明Description of the drawings
图1是本申请实施例一提供的视频分析方法的流程图。FIG. 1 is a flowchart of a video analysis method provided in Embodiment 1 of the present application.
图2是本申请实施例二提供的本申请视频分析装置较佳实施例中的功能模块图。2 is a diagram of functional modules in a preferred embodiment of the video analysis device of this application provided in the second embodiment of this application.
图3是本申请实施例三提供的计算机设备的示意图。Fig. 3 is a schematic diagram of a computer device provided in Embodiment 3 of the present application.
如下具体实施方式将结合上述附图进一步说明本申请。The following specific embodiments will further illustrate this application in conjunction with the above-mentioned drawings.
具体实施方式Detailed ways
为了能够更清楚地理解本申请的上述目的、特征和优点,下面结合附图和具体实施例对本申请进行详细描述。需要说明的是,在不冲突的情况下,本申请的实施例及实施例中的特征可以相互组合。In order to be able to understand the above objectives, features and advantages of the application more clearly, the application will be described in detail below with reference to the accompanying drawings and specific embodiments. It should be noted that the embodiments of the application and the features in the embodiments can be combined with each other if there is no conflict.
在下面的描述中阐述了很多具体细节以便于充分理解本申请,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In the following description, many specific details are set forth in order to fully understand the present application. The described embodiments are only a part of the embodiments of the present application, rather than all the embodiments. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of this application.
除非另有定义,本文所使用的所有的技术和科学术语与属于本申请的技术领域的技术人员通常理解的含义相同。本文中在本申请的说明书中所使用的术语只是为了描述具体的实施例的目的,不是旨在于限制本申请。Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by those skilled in the technical field of this application. The terms used in the description of the application herein are only for the purpose of describing specific embodiments, and are not intended to limit the application.
本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”和“第三”等是用于区别不同对象,而非用于描述特定顺序。此外,术语“包括”以及它们任何变形,意图在于覆盖不排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或设备没有限定于已列出的步骤或单元,而是可选地还包括没有列出的步骤或单元,或可选地还包括对于这些过程、方法、产品或设备固有的其它步骤或单元。The terms "first", "second", and "third" in the specification and claims of this application and the above-mentioned drawings are used to distinguish different objects, rather than to describe a specific sequence. In addition, the term "including" and any variations of them are intended to cover non-exclusive inclusion. For example, a process, method, system, product, or device that includes a series of steps or units is not limited to the listed steps or units, but optionally includes unlisted steps or units, or optionally also includes Other steps or units inherent to these processes, methods, products or equipment.
本申请实施例的视频分析方法应用在由至少一个计算机设备和通过网络与所述计算机设备进行连接的移动终端所构成的硬件环境中。网络包括但不限于:广域网、城域网或局域网。本申请实施例的视频分析方法可以由计算机设备来执行,也可以由移动终端来执行;还可以是由计算机设备和移动终 端共同执行。The video analysis method of the embodiment of the present application is applied in a hardware environment composed of at least one computer device and a mobile terminal connected to the computer device through a network. Networks include but are not limited to: wide area network, metropolitan area network or local area network. The video analysis method in the embodiments of the present application may be executed by a computer device or a mobile terminal; it may also be executed by the computer device and the mobile terminal.
所述对于需要进行视频分析方法的计算机设备,可以直接在计算机设备上集成本申请的方法所提供的视频分析功能,或者安装用于实现本申请的方法的客户端。再如,本申请所提供的方法还可以以软件开发工具包(Software Development Kit,SDK)的形式运行在计算机等设备上,以SDK的形式提供视频分析功能的接口,计算机或其他设备通过提供的接口即可实现视频分析功能。For the computer equipment that needs to perform the video analysis method, the video analysis function provided by the method of this application can be directly integrated on the computer device, or a client for implementing the method of this application can be installed. For another example, the method provided in this application can also be run on a computer or other device in the form of a software development kit (SDK), and provide an interface for video analysis functions in the form of SDK. The interface can realize the video analysis function.
实施例一Example one
图1是本申请实施例一提供的视频分析方法的流程图。根据不同的需求,该流程图中的执行顺序可以改变,某些步骤可以省略。FIG. 1 is a flowchart of a video analysis method provided in Embodiment 1 of the present application. According to different needs, the execution order in this flowchart can be changed, and some steps can be omitted.
步骤S1,接收摄像头采集的视频图像。Step S1, receiving the video image collected by the camera.
在本实施方式中,通过摄像头采集视频图像,所述摄像头被安装在不同的业务场景中。所述业务场景描述的是需要进行目标对象侦测和/或视频分析的场景。例如,所述业务场景为识别交通事故、拥堵、车速检测、车流预测、车辆失控、车辆行驶轨迹、人员或自行车闯入、违反交通法规、抛洒物等的智能交通业务场景,所述业务场景还可以是识别人员入侵、遗留物、遗失物监测、车牌分析、车辆行驶轨迹、车流分析、人流分析、烟火或烟雾等的智慧园区业务场景,所述业务场景还可以是非法船只、超载、密集人群检测、是否穿救生衣、落水等的渡口监测业务场景。In this embodiment, the video image is collected by a camera, and the camera is installed in different business scenarios. The business scenario describes a scenario that requires target object detection and/or video analysis. For example, the business scenario is an intelligent traffic business scenario that recognizes traffic accidents, congestion, vehicle speed detection, traffic flow prediction, vehicle loss of control, vehicle trajectory, people or bicycle intrusion, violation of traffic laws, throwing objects, etc. The business scenario also It can be a smart park business scenario that identifies personnel intrusion, leftovers, lost property monitoring, license plate analysis, vehicle trajectory, traffic flow analysis, pedestrian flow analysis, fireworks or smoke, etc. The business scenario can also be illegal ships, overloaded, dense crowds, etc. Ferry monitoring business scenarios such as detection, whether to wear a life jacket, falling into the water, etc.
所述业务场景还可以是无人驾驶、金融场景、设备登录、机场及公共区域的监控等场景。The business scenarios may also be scenarios such as unmanned driving, financial scenarios, equipment login, airport and public area monitoring.
在本实施方式中,所述摄像头可以是不同厂商出厂的不同型号规格的摄像头,所述视频分析方法可以实现统一处理并分析不同厂商出厂的不同型号规格的摄像头拍摄的视频图像。In this embodiment, the cameras may be cameras of different models and specifications manufactured by different manufacturers, and the video analysis method can realize unified processing and analysis of video images taken by cameras of different models and specifications manufactured by different manufacturers.
优选地,接收摄像头采集的视频图像后,所述视频分析方法还包括:Preferably, after receiving the video image collected by the camera, the video analysis method further includes:
对所述视频图像进行解码的步骤。The step of decoding the video image.
具体地,可以通过图形处理器(GPU)对所述视频图像进行视频解码,以得到所述视频图像中的每帧图像。Specifically, the video image may be video decoded by a graphics processing unit (GPU) to obtain each frame of the video image.
步骤S2,检测所述视频图像中的目标对象得到所述目标对象的类别。Step S2: Detect the target object in the video image to obtain the target object category.
在本实施方式中,所述视频图像中的目标对象包括人物、动物、交通工具、建筑物、烟雾等。In this embodiment, the target objects in the video image include people, animals, vehicles, buildings, smoke and so on.
具体地,所述检测所述视频图像中的目标对象得到所述目标对象的类别包括:Specifically, detecting the target object in the video image to obtain the target object category includes:
(1)识别所述视频图像中的目标对象;(1) Identify the target object in the video image;
在本实施方式中,所述视频图像中的目标对象包括静止目标对象和运动目标对象。In this embodiment, the target object in the video image includes a static target object and a moving target object.
当所述视频图像中的目标对象为静止目标对象时,可以通过基于模板的检测方法来识别所述静止目标对象。具体包括:确定所述视频图像中的目标对象形状的轮廓,将所述目标对象形状的轮廓与预存的模板文件进行特征匹 配。When the target object in the video image is a stationary target object, the stationary target object can be identified through a template-based detection method. Specifically, it includes: determining the contour of the target object shape in the video image, and matching the contour of the target object shape with a pre-stored template file.
例如,当所述视频图像中的目标对象为一扇门,可以确定所述目标对象形状的轮廓为一矩形,将所述矩形与预存的门的模板文件进行特征匹配来识别所述目标对象。其中,所述门的模板文件为矩形。For example, when the target object in the video image is a door, the outline of the shape of the target object can be determined to be a rectangle, and the rectangle is feature-matched with a pre-stored door template file to identify the target object. Wherein, the template file of the door is rectangular.
当所述视频图像中的目标对象为运动目标对象时,可以通过背景差法、帧差法、光流法中的至少一种进行识别。所述背景差法是对视频图像中相对较为固定的场景进行背景建模,检测时由当前图像与背景模型之差得到所述运动目标对象;所述帧差法是通过对视频序列中相邻帧之间对应位置像素点进行比较来获取运动目标对象的位置;所述光流法是利用时间变化的光流矢量特性,对所述视频图像中的运动目标对象进行检测。When the target object in the video image is a moving target object, it can be identified by at least one of the background difference method, the frame difference method, and the optical flow method. The background difference method is to perform background modeling on a relatively fixed scene in the video image, and the moving target object is obtained from the difference between the current image and the background model during detection; The corresponding position pixels between the frames are compared to obtain the position of the moving target object; the optical flow method uses the time-varying optical flow vector characteristics to detect the moving target object in the video image.
可以理解的是,上述检测视频图像中的静止目标对象和运动目标对象的方法不限于上述列举的,任何适应于检测出视频图像中的目标对象的放法均可应用于此。另外,本实施例中的所述检测视频图像中的静止目标对象和运动目标对象的方法均为现有技术,本文在此不再详细介绍。It is understandable that the above-mentioned method for detecting static target objects and moving target objects in a video image is not limited to the above-mentioned enumeration, and any reproduction method suitable for detecting a target object in a video image can be applied to this. In addition, the methods for detecting stationary target objects and moving target objects in a video image in this embodiment are all existing technologies, and will not be described in detail herein.
(2)确定所述目标对象的类别。(2) Determine the category of the target object.
例如,当识别所述视频图像中的目标对象为汽车时,可以确定所述目标对象的类别为交通工具。所述目标对象的检测和分类是视觉技术中一个非常基础的任务,其目的就是跟踪场景中感兴趣的一些物体,包括常规的目标对象检测、人员检测以及车辆检测等等。For example, when the target object in the video image is recognized as a car, it may be determined that the target object is a vehicle. The detection and classification of the target object is a very basic task in vision technology, and its purpose is to track some objects of interest in the scene, including conventional target object detection, person detection, vehicle detection, and so on.
在本实施方式中,可以通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性,其中,所述基本属性包括颜色、运动轨迹、形状、结构等,再在所述获取的基本属性与预先存储在数据库中的目标对象的基本属性进行比对,从而准确地识别出所述视频图像中的目标对象。所述数据库中存储有目标对象的基本属性与目标对象类别对应表。In this embodiment, the basic attributes of the target object in the video image can be obtained by decomposing the target object in the video image, where the basic attributes include color, motion track, shape, structure, etc., and then The obtained basic attributes are compared with the basic attributes of the target object pre-stored in the database, so as to accurately identify the target object in the video image. The database stores a table corresponding to the basic attributes of the target object and the target object category.
所述确定所述目标对象的类别具体包括:通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性;将获取的所述基本属性与预先存储在数据库中的目标对象的基本属性进行比对;当获取的所述基本属性与所述数据库中的目标对象的基本属性一致时,查询数据库中存储的基本属性与目标对象类别对应表以得到所述目标对象的类别。The determining the category of the target object specifically includes: obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image; and comparing the obtained basic attributes with the ones stored in a database in advance. The basic attributes of the target object are compared; when the acquired basic attributes are consistent with the basic attributes of the target object in the database, the corresponding table of basic attributes and target object categories stored in the database is queried to obtain the target object’s category.
步骤S3,跟踪所述视频图像中的目标对象得到所述目标对象的状态。Step S3, tracking the target object in the video image to obtain the state of the target object.
在目标对象检测完成以后,需要针对每个检测到的目标对象来计算其运动轨迹,从而实现跟踪所述视频图像中的目标对象。在本实施方式中,通过跟踪所述视频图像中的目标对象可以确定所述目标对象的状态。After the target object detection is completed, it is necessary to calculate the motion trajectory of each detected target object, so as to realize the tracking of the target object in the video image. In this embodiment, the state of the target object can be determined by tracking the target object in the video image.
所述跟踪所述视频图像中的目标对象的方法包括:The method for tracking the target object in the video image includes:
a)确定当前视频帧中的目标对象。a) Determine the target object in the current video frame.
b)获取目标对象在前序视频帧中的图像区域以及所述图像区域的图像特征,其中,所述前序视频帧为当前视频帧之前的k个视频帧,k为正整数。b) Obtain the image area of the target object in the previous video frame and the image characteristics of the image area, wherein the previous video frame is k video frames before the current video frame, and k is a positive integer.
c)根据所述目标对象在前序视频帧中的图像区域,对所述目标对象进行运动估计,确定所述目标对象在当前视频帧的预测区域。c) Perform motion estimation on the target object according to the image area of the target object in the previous video frame, and determine the prediction area of the target object in the current video frame.
d)根据所述预测区域确定目标对象在当前视频帧中的检测范围。d) Determine the detection range of the target object in the current video frame according to the prediction area.
e)判断所述目标对象是否出现在当前视频帧中的检测范围,若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象的状态异常;若所述目标对象出现在当前视频帧中的检测范围,确定所述目标对象在当前视频帧中的图像区域,即所述目标对象的状态正常。e) Determine whether the target object appears in the detection range of the current video frame, if the target object does not appear in the detection range of the current video frame, determine that the state of the target object is abnormal; if the target object appears in the detection range The detection range in the current video frame determines the image area of the target object in the current video frame, that is, the state of the target object is normal.
由于前序视频帧是指当前视频帧之前的k个视频帧,通过这前k个视频帧来对当前视频帧进行预估和对比检测,计算量较小,并且能够解决视频中目标对象偶尔丢失或者遮挡的问题,检测精度较高。Since the pre-order video frame refers to the k video frames before the current video frame, the current video frame is estimated and compared and detected through the first k video frames, which requires a small amount of calculation and can solve the occasional loss of target objects in the video Or the problem of occlusion, the detection accuracy is higher.
步骤S4,根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景。Step S4: Analyze the business scene contained in the video image according to the category of the target object and the state of the target object.
在本实施方式中,根据检测结果可以获得目标对象的类别,根据跟踪结果可以确定所述目标对象的状态,从而可以分析得到所述视频图像中包含的业务场景。In this embodiment, the category of the target object can be obtained according to the detection result, and the state of the target object can be determined according to the tracking result, so that the business scene contained in the video image can be analyzed.
例如,根据检测结果可以获得所述目标对象的类别为汽车,所述汽车没有出现在当前视频帧中的检测范围,则可以确定所述汽车的状态异常,如所述汽车处于拥堵状态,则可得知所述视频图像中包括的业务场景为智能交通业务场景。For example, according to the detection result, it can be obtained that the category of the target object is a car. If the car does not appear in the detection range of the current video frame, it can be determined that the state of the car is abnormal. If the car is in a congested state, it can be It is learned that the business scene included in the video image is an intelligent transportation business scene.
又如根据检测结果可以获得所述目标对象的类别为行人,所述行人没有出现在当前视频帧中的检测范围,则可以确定所述行人的状态异常。如所述行人摔倒,则可得知所述视频图像中包括的业务场景为智能交通业务场景。In another example, according to the detection result, the category of the target object can be obtained as a pedestrian, and the pedestrian does not appear in the detection range in the current video frame, it can be determined that the state of the pedestrian is abnormal. If the pedestrian falls down, it can be known that the business scene included in the video image is an intelligent traffic business scene.
又如根据检测结果可以获得所述目标对象的类别为门,所述门没有出现在当前视频中的检测范围,则可以确认所述门的状态异常,如保持打开状态可判断所述视频图像中包括的业务场景为智能安保业务场景。Also, if the category of the target object can be obtained as a door according to the detection result, and the door does not appear in the detection range in the current video, it can be confirmed that the status of the door is abnormal. If it is kept open, it can be judged that the door is in the video image. The business scenarios included are smart security business scenarios.
步骤S5,判断所述视频图像中的业务场景是否出现异常。当所述视频图像中的业务场景出现异常时,进入步骤S6;当所述视频图像中的业务场景没有出现异常时,结束流程。Step S5: Determine whether the business scene in the video image is abnormal. When the business scene in the video image is abnormal, step S6 is entered; when the business scene in the video image is not abnormal, the process ends.
在本实施方式中,通过所述目标对象的类别和所述目标对象的状态分析可判断所述视频图像中的业务场景是否出现异常。例如,通过判断所述目标对象是否出现在当前视频帧中的检测范围,若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象的状态异常,即所述目标对象对应的业务场景也出现异常。In this embodiment, it can be determined whether the business scene in the video image is abnormal by analyzing the category of the target object and the state of the target object. For example, by determining whether the target object appears in the detection range of the current video frame, if the target object does not appear in the detection range of the current video frame, it is determined that the state of the target object is abnormal, that is, the target object corresponds to The business scenario also appeared abnormal.
在其他实施方式中,可以通过将所述视频图像输入预先训练好的异常模型,并根据所述异常模型判断所述视频图像中的业务场景是否异常。具体地,当确定所述目标对象异常时,提取所述当前视频帧作为异常图像;将所述异常图像作为待识别图像导入预先训练好的异常模型中,其中,所述异常模型用于表征待识别图像与异常场景之间的对应关系;当所述异常模型输出与所述待识别图像对应的异常场景时,确认所述业务场景出现异常所述异常模型包括不同的业务场景对应的异常模型。例如,当所述业务场景为智能交通业务场景时,所述智能交通业务场景对应的异常模型包括交通事故模型、交通 拥堵模型及违法违规模型等;当所述业务场景为智慧园区业务场景时,所述智慧园区业务场景对应的异常模型包括随身物品遗留模型、人员入侵模型等;当所述业务场景为渡口监测业务场景时,所述渡口监测业务场景对应的异常模型包括超载模型、落水模型、非法船只模型等。In other implementation manners, the video image may be input to a pre-trained abnormality model, and whether the business scene in the video image is abnormal or not can be determined according to the abnormality model. Specifically, when it is determined that the target object is abnormal, extract the current video frame as an abnormal image; import the abnormal image as an image to be recognized into a pre-trained anomaly model, wherein the abnormal model is used to characterize the Identify the correspondence between the image and the abnormal scene; when the abnormal model outputs the abnormal scene corresponding to the image to be identified, confirm that the business scene is abnormal and the abnormal model includes abnormal models corresponding to different business scenarios. For example, when the business scene is an intelligent transportation business scene, the abnormal model corresponding to the intelligent transportation business scene includes a traffic accident model, a traffic congestion model, and an illegal scale type, etc.; when the business scene is a smart park business scene, The abnormal model corresponding to the business scene of the smart park includes a personal belongings model, a personnel intrusion model, etc.; when the business scene is a ferry monitoring business scene, the abnormal model corresponding to the ferry monitoring business scene includes an overload model, a falling water model, Illegal ship model, etc.
举例而言,当所述视频图像中当前视频帧出现交通拥堵时,提取所述当前视频帧作为异常图像,并将所述异常图像作为待识别图像导入预先训练好的交通拥堵模型中,当所述交通拥堵模型输出与所述待识别图像对应的交通拥堵场景时,确认所述视频图像对应的智能交通业务场景中出现异常;当所述交通拥堵模型没有输出与所述待识别图像对应的交通拥堵场景时,确认所述视频图像对应的智能交通业务场景中正常。For example, when traffic congestion occurs in the current video frame in the video image, the current video frame is extracted as an abnormal image, and the abnormal image is imported into a pre-trained traffic congestion model as an image to be recognized. When the traffic jam model outputs the traffic jam scene corresponding to the image to be recognized, it is confirmed that there is an abnormality in the intelligent transportation business scene corresponding to the video image; when the traffic jam model does not output the traffic corresponding to the image to be recognized In a congested scene, confirm that the intelligent transportation service scene corresponding to the video image is normal.
上述异常模型为根据图片样本集训练的机器学习模型。所述图片样本包括异常业务场景图片样本和正常业务场景图片样本。所述机器学习模型为可以进行图像识别的人工智能算法模型,包括:卷积神经网络模型CNN、循环神经网络模块RNN和深度神经网络模型DNN。其中,卷积神经网络模型CNN是一种多层神经网络,可以将数据量庞大的图像识别问题不断降维,最终使其能够被训练,因此,本申请实施例中的机器学习模型可以为CNN模型。The above-mentioned abnormal model is a machine learning model trained based on a picture sample set. The picture samples include abnormal business scene picture samples and normal business scene picture samples. The machine learning model is an artificial intelligence algorithm model that can perform image recognition, including: a convolutional neural network model CNN, a recurrent neural network module RNN, and a deep neural network model DNN. Among them, the convolutional neural network model CNN is a multi-layer neural network, which can continuously reduce the dimensionality of the image recognition problem with a huge amount of data, and finally enable it to be trained. Therefore, the machine learning model in the embodiment of the present application may be CNN model.
在CNN网络结构的演化上,出现了许多CNN网络,包括LeNet、AlexNet、VGGNet、GoogleNet和ResNet。其中,ResNet网络提出了一种减轻网络训练负担的残差学习框架,这种网络比以前使用过的网络本质上层次更深,解决了其他神经网络随着网络加深,准确率下降的问题。在本实施方式中,所述机器学习模型可以是卷积申请网络模型CNN中的ResNet模型。需要说明的是,此处仅是举例说明,其他可以进行图像识别的机器学习模型同样适用于本申请,此处不进行赘述。In the evolution of CNN network structure, many CNN networks have appeared, including LeNet, AlexNet, VGGNet, GoogleNet and ResNet. Among them, the ResNet network proposes a residual learning framework that reduces the burden of network training. This network is inherently deeper than the previously used network, and solves the problem of other neural networks that decrease in accuracy as the network deepens. In this embodiment, the machine learning model may be the ResNet model in the convolution application network model CNN. It should be noted that this is only an example, and other machine learning models that can perform image recognition are also applicable to this application, and will not be repeated here.
步骤S6,当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。Step S6: When the business scene in the video image is abnormal, record the key information when the business scene is abnormal.
在本实施方式中,所述关键信息包括所述业务场景出现异常的时间、地点、及截取的所述视频图像中所述业务场景出现异常时的图片文件等。In this embodiment, the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
进一步地,所述视频分析方法还包括,将记录的关键信息发送至第三方业务平台。所述第三方业务平台包括公安系统、交通管制系统等。通过将所述记录的关键信息发送至所述第三方业务平台,可以帮助所述第三方业务平台及时获取业务场景中出现异常时的关键信息,从而及时处理所述异常。Further, the video analysis method further includes sending the recorded key information to a third-party service platform. The third-party service platform includes a public security system, a traffic control system, etc. By sending the recorded key information to the third-party business platform, the third-party business platform can help the third-party business platform to obtain key information when an exception occurs in a business scenario in time, so as to process the exception in time.
进一步地,所述视频分析方法还包括:展示所述业务场景出现异常时的关键信息。具体地,在所述显示屏中显示所述业务场景的异常图片、时间、点的等信息。Further, the video analysis method further includes: displaying key information when the business scenario is abnormal. Specifically, information such as an abnormal picture, time, and point of the business scene is displayed on the display screen.
综上所述,本申请提供的视频分析方法,包括接收摄像头采集的视频图像;检测所述视频图像中的目标对象得到所述目标对象的类别;跟踪所述视频图像中的目标对象得到所述目标对象的状态;根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;判断所述业务场景是否出现异常;及当所述视频图像中的业务场景出现异常时,记录所 述业务场景出现异常时的关键信息。可以实时分析所述视频图像对应的业务场景是否出现异常,并在确认所述业务场景出现异常时,记录所述业务场景出现异常时的关键信息。从而可以将所述关键信息发送至对应的第三方平台,以及时处理所述异常。In summary, the video analysis method provided by this application includes receiving a video image collected by a camera; detecting a target object in the video image to obtain the category of the target object; tracking the target object in the video image to obtain the The status of the target object; the business scene contained in the video image is obtained by analyzing the category of the target object and the status of the target object; judging whether the business scene is abnormal; and when the business scene in the video image When an exception occurs, record the key information when the business scenario is abnormal. It is possible to analyze in real time whether the business scene corresponding to the video image is abnormal, and when it is confirmed that the business scene is abnormal, record the key information when the business scene is abnormal. Thereby, the key information can be sent to the corresponding third-party platform, and the exception can be handled in time.
实施例二Example two
图2为本申请视频分析装置较佳实施例中的功能模块图。FIG. 2 is a diagram of functional modules in a preferred embodiment of the video analysis device of this application.
在一些实施例中,所述视频分析装置20运行于计算机设备中。所述视频分析装置20可以包括多个由计算机可读指令代码段所组成的功能模块。所述视频分析装置20中的各个计算机可读指令代码段的指令代码可以存储于存储器中,并由至少一个处理器所执行,以执行(详见图1及其相关描述)视频分析功能。In some embodiments, the video analysis device 20 runs in a computer device. The video analysis device 20 may include multiple functional modules composed of computer-readable instruction code segments. The instruction codes of each computer-readable instruction code segment in the video analysis device 20 can be stored in a memory and executed by at least one processor to perform (see FIG. 1 and related descriptions for details) video analysis functions.
本实施例中,所述视频分析装置20根据其所执行的功能,可以被划分为多个功能模块。所述功能模块可以包括:接收模块201、检测模块202、跟踪模块203、分析模块204、判断模块205及处理模块206。本申请所称的模块是指一种能够被至少一个处理器所执行并且能够完成固定功能的一系列计算机可读指令代码段,其存储在存储器中。在一些实施例中,关于各模块的功能将在后续的实施例中详述。In this embodiment, the video analysis device 20 can be divided into multiple functional modules according to the functions it performs. The functional modules may include: a receiving module 201, a detection module 202, a tracking module 203, an analysis module 204, a judgment module 205, and a processing module 206. The module referred to in this application refers to a series of computer-readable instruction code segments that can be executed by at least one processor and can complete fixed functions, and are stored in a memory. In some embodiments, the functions of each module will be detailed in subsequent embodiments.
所述接收模块201用于接收摄像头采集的视频图像。The receiving module 201 is used to receive video images collected by a camera.
在本实施方式中,通过摄像头采集视频图像,所述摄像头被安装在不同的业务场景中。所述业务场景描述的是需要进行目标对象侦测和/或视频分析的场景。例如,所述业务场景为识别交通事故、拥堵、车速检测、车流预测、车辆失控、车辆行驶轨迹、人员或自行车闯入、违反交通法规、抛洒物等的智能交通业务场景,所述业务场景还可以是识别人员入侵、遗留物、遗失物监测、车牌分析、车辆行驶轨迹、车流分析、人流分析、烟火或烟雾等的智慧园区业务场景,所述业务场景还可以是非法船只、超载、密集人群检测、是否穿救生衣、落水等的渡口监测业务场景。In this embodiment, the video image is collected by a camera, and the camera is installed in different business scenarios. The business scenario describes a scenario that requires target object detection and/or video analysis. For example, the business scenario is an intelligent traffic business scenario that recognizes traffic accidents, congestion, vehicle speed detection, traffic flow prediction, vehicle loss of control, vehicle trajectory, people or bicycle intrusion, violation of traffic laws, throwing objects, etc. The business scenario also It can be a smart park business scenario that identifies personnel intrusion, leftovers, lost property monitoring, license plate analysis, vehicle trajectory, traffic flow analysis, pedestrian flow analysis, fireworks or smoke, etc. The business scenario can also be illegal ships, overloaded, dense crowds, etc. Ferry monitoring business scenarios such as detection, whether to wear a life jacket, falling into the water, etc.
在本实施方式中,所述摄像头与所述计算机设备之间通过有线或无线网络通信连接。所述摄像头将采集的视频图像通过有线或无线网络发送至所述计算机设备。In this embodiment, the camera and the computer device are connected through a wired or wireless network communication. The camera sends the collected video images to the computer device through a wired or wireless network.
所述业务场景还可以是无人驾驶、金融场景、设备登录、机场及公共区域的监控等场景。The business scenarios may also be scenarios such as unmanned driving, financial scenarios, equipment login, airport and public area monitoring.
在本实施方式中,所述摄像头可以是不同厂商出厂的不同型号规格的摄像头,所述所述视频分析装置20可以实现统一处理并分析不同厂商出厂的不同型号规格的摄像头拍摄的视频图像。In this embodiment, the cameras may be cameras of different models and specifications manufactured by different manufacturers, and the video analysis device 20 may implement unified processing and analysis of video images taken by cameras of different models and specifications manufactured by different manufacturers.
优选地,接收摄像头采集的视频图像后,所述视频分析装置20还可以对所述视频图像进行解码。Preferably, after receiving the video image collected by the camera, the video analysis device 20 may also decode the video image.
具体地,可以通过图形处理器(GPU)对所述视频图像进行视频解码,以得到所述视频图像中的每帧图像。Specifically, the video image may be video decoded by a graphics processing unit (GPU) to obtain each frame of the video image.
所述检测模块202用于检测所述视频图像中的目标对象得到所述目标对 象的类别。The detection module 202 is used to detect the target object in the video image to obtain the target object category.
在本实施方式中,所述视频图像中的目标对象包括人物、动物、交通工具、建筑物、烟雾等。In this embodiment, the target objects in the video image include people, animals, vehicles, buildings, smoke and so on.
具体地,所述检测所述视频图像中的目标对象得到所述目标对象的类别包括:Specifically, detecting the target object in the video image to obtain the target object category includes:
(1)识别所述视频图像中的目标对象;(1) Identify the target object in the video image;
在本实施方式中,所述视频图像中的目标对象包括静止目标对象和运动目标对象。In this embodiment, the target object in the video image includes a static target object and a moving target object.
当所述视频图像中的目标对象为静止目标对象时,可以通过基于模板的检测方法来识别所述静止目标对象。具体包括:确定所述视频图像中的目标对象形状的轮廓,将所述目标对象形状的轮廓与预存的模板文件进行特征匹配。When the target object in the video image is a stationary target object, the stationary target object can be identified through a template-based detection method. Specifically, it includes: determining the contour of the target object shape in the video image, and performing feature matching between the contour of the target object shape and a pre-stored template file.
当所述视频图像中的目标对象为运动目标对象时,可以通过背景差法、帧差法、光流法中的至少一种进行识别。所述背景差法是对视频图像中相对较为固定的场景进行背景建模,检测时由当前图像与背景模型之差得到所述运动目标对象;所述帧差法是通过对视频序列中相邻帧之间对应位置像素点进行比较来获取运动目标对象的位置;所述光流法是利用时间变化的光流矢量特性,对所述视频图像中的运动目标对象进行检测。When the target object in the video image is a moving target object, it can be identified by at least one of the background difference method, the frame difference method, and the optical flow method. The background difference method is to perform background modeling on a relatively fixed scene in the video image, and the moving target object is obtained from the difference between the current image and the background model during detection; the frame difference method is to obtain the moving target object from the difference between the current image and the background model; The corresponding position pixels between the frames are compared to obtain the position of the moving target object; the optical flow method uses the time-varying optical flow vector characteristics to detect the moving target object in the video image.
在本实施方式中,上述检测视频图像中的静止目标对象和运动目标对象的方法不限于上述列举的,任何适应于检测出视频图像中的目标对象的放法均可应用于此。另外,本实施例中的所述检测视频图像中的静止目标对象和运动目标对象的方法均为现有技术,本文在此不再详细介绍。In this embodiment, the method for detecting the stationary target object and the moving target object in the video image is not limited to the above-listed method, and any reproduction method suitable for detecting the target object in the video image can be applied to this. In addition, the methods for detecting stationary target objects and moving target objects in a video image in this embodiment are all existing technologies, and will not be described in detail herein.
(2)确定所述目标对象的类别。(2) Determine the category of the target object.
例如,当识别所述视频图像中的目标对象为汽车时,可以确定所述目标对象的类别为交通工具。所述目标对象的检测和分类是视觉技术中一个非常基础的任务,其目的就是跟踪场景中感兴趣的一些物体,包括常规的目标对象检测、人员检测以及车辆检测等等。For example, when the target object in the video image is recognized as a car, it may be determined that the target object is a vehicle. The detection and classification of the target object is a very basic task in vision technology, and its purpose is to track some objects of interest in the scene, including conventional target object detection, person detection, vehicle detection, and so on.
在本实施方式中,可以通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性,其中,所述基本属性包括颜色、运动轨迹、形状、结构等,再在所述获取的基本属性与预先存储在数据库中的目标对象的基本属性进行比对,从而准确地识别出所述视频图像中的目标对象。所述数据库中存储有目标对象的基本属性与目标对象类别对应表。In this embodiment, the basic attributes of the target object in the video image can be obtained by decomposing the target object in the video image, where the basic attributes include color, motion track, shape, structure, etc., and then The obtained basic attributes are compared with the basic attributes of the target object pre-stored in the database, so as to accurately identify the target object in the video image. The database stores a table corresponding to the basic attributes of the target object and the target object category.
所述确定所述目标对象的类别具体包括:通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性;将获取的所述基本属性与预先存储在数据库中的目标对象的基本属性进行比对;当获取的所述基本属性与所述数据库中的目标对象的基本属性一致时,查询数据库中存储的基本属性与目标对象类别对应表以得到所述目标对象的类别。The determining the category of the target object specifically includes: obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image; and comparing the obtained basic attributes with the ones stored in a database in advance. The basic attributes of the target object are compared; when the acquired basic attributes are consistent with the basic attributes of the target object in the database, the corresponding table of basic attributes and target object categories stored in the database is queried to obtain the target object’s category.
所述跟踪模块203用于跟踪所述视频图像中的目标对象得到所述目标对象的状态。The tracking module 203 is configured to track the target object in the video image to obtain the state of the target object.
在目标对象检测完成以后,需要针对每个检测到的目标对象来计算其运动轨迹,从而实现跟踪所述视频图像中的目标对象。在本实施方式中,通过跟踪所述视频图像中的目标对象可以确定所述目标对象的状态。After the target object detection is completed, it is necessary to calculate the motion trajectory of each detected target object, so as to realize the tracking of the target object in the video image. In this embodiment, the state of the target object can be determined by tracking the target object in the video image.
所述跟踪所述视频图像中的目标对象的方法包括:The method for tracking the target object in the video image includes:
a)确定当前视频帧中的目标对象。a) Determine the target object in the current video frame.
b)获取目标对象在前序视频帧中的图像区域以及所述图像区域的图像特征,其中,所述前序视频帧为当前视频帧之前的k个视频帧,k为正整数。b) Obtain the image area of the target object in the previous video frame and the image characteristics of the image area, wherein the previous video frame is k video frames before the current video frame, and k is a positive integer.
c)根据所述目标对象在前序视频帧中的图像区域,对所述目标对象进行运动估计,确定所述目标对象在当前视频帧的预测区域。c) Perform motion estimation on the target object according to the image area of the target object in the previous video frame, and determine the prediction area of the target object in the current video frame.
d)根据所述预测区域确定目标对象在当前视频帧中的检测范围。d) Determine the detection range of the target object in the current video frame according to the prediction area.
e)判断所述目标对象是否出现在当前视频帧中的检测范围,若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象的状态异常;若所述目标对象出现在当前视频帧中的检测范围,确定所述目标对象在当前视频帧中的图像区域。e) Determine whether the target object appears in the detection range of the current video frame, if the target object does not appear in the detection range of the current video frame, determine that the state of the target object is abnormal; if the target object appears in the detection range The detection range in the current video frame determines the image area of the target object in the current video frame.
由于前序视频帧是指当前视频帧之前的k个视频帧,通过这前k个视频帧来对当前视频帧进行预估和对比检测,计算量较小,并且能够解决视频中目标对象偶尔丢失或者遮挡的问题,检测精度较高。Since the pre-order video frame refers to the k video frames before the current video frame, the current video frame is estimated and compared and detected through the first k video frames, which requires a small amount of calculation and can solve the occasional loss of target objects in the video Or the problem of occlusion, the detection accuracy is higher.
所述分析模块204用于根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景。The analysis module 204 is configured to analyze and obtain the business scenario contained in the video image according to the category of the target object and the state of the target object.
在本实施方式中,根据检测结果可以获得目标对象的类别,根据跟踪结果可以确定所述目标对象的状态,从而可以分析得到所述视频图像中包含的业务场景。In this embodiment, the category of the target object can be obtained according to the detection result, and the state of the target object can be determined according to the tracking result, so that the business scene contained in the video image can be analyzed.
例如,根据检测结果可以获得所述目标对象为汽车,所述汽车没有出现在当前视频帧中的检测范围,则可以确定所述汽车的状态异常,如所述汽车处于拥堵状态,则可得知所述视频图像中包括的业务场景为智能交通业务场景。For example, according to the detection result, it can be obtained that the target object is a car. If the car does not appear in the detection range of the current video frame, it can be determined that the state of the car is abnormal. If the car is in a congested state, it can be known The business scene included in the video image is an intelligent transportation business scene.
又如根据检测结果可以获得所述目标对象的类别为行人,所述行人没有出现在当前视频帧中的检测范围,则可以确定所述行人的状态异常。如所述行人摔倒,则可得知所述视频图像中包括的业务场景为智能交通业务场景。In another example, according to the detection result, the category of the target object can be obtained as a pedestrian, and the pedestrian does not appear in the detection range in the current video frame, it can be determined that the state of the pedestrian is abnormal. If the pedestrian falls down, it can be known that the business scene included in the video image is an intelligent traffic business scene.
所述判断模块205用于判断所述视频图像中的业务场景是否出现异常。当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。The judgment module 205 is used to judge whether the business scene in the video image is abnormal. When the business scene in the video image is abnormal, the key information when the business scene is abnormal is recorded.
在本实施方式中,通过所述目标对象的类别和所述目标对象的状态分析可判断所述视频图像中的业务场景是否出现异常。例如,通过判断所述目标对象是否出现在当前视频帧中的检测范围,若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象的状态异常,即所述目标对象对应的业务场景也出现异常。In this embodiment, it can be determined whether the business scene in the video image is abnormal by analyzing the category of the target object and the state of the target object. For example, by determining whether the target object appears in the detection range of the current video frame, if the target object does not appear in the detection range of the current video frame, it is determined that the state of the target object is abnormal, that is, the target object corresponds to The business scenario also appeared abnormal.
在其他实施方式中,可以通过将所述视频图像输入预先训练好的异常模型,并根据所述异常模型判断所述视频图像中的业务场景是否异常。具体地, 当确定所述目标对象异常时,提取所述当前视频帧作为异常图像;将所述异常图像作为待识别图像导入预先训练好的异常模型中,其中,所述异常模型用于表征待识别图像与异常场景之间的对应关系;当所述异常模型输出与所述待识别图像对应的异常场景时,确认所述业务场景出现异常所述异常模型包括不同的业务场景对应的异常模型。例如,当所述业务场景为智能交通业务场景时,所述智能交通业务场景对应的异常模型包括交通事故模型、交通拥堵模型及违法违规模型等;当所述业务场景为智慧园区业务场景时,所述智慧园区业务场景对应的异常模型包括随身物品遗留模型、人员入侵模型等;当所述业务场景为渡口监测业务场景时,所述渡口监测业务场景对应的异常模型包括超载模型、落水模型、非法船只模型等。In other implementation manners, the video image may be input to a pre-trained abnormality model, and whether the business scene in the video image is abnormal or not can be determined according to the abnormality model. Specifically, when it is determined that the target object is abnormal, extract the current video frame as an abnormal image; import the abnormal image as an image to be recognized into a pre-trained abnormal model, where the abnormal model is used to characterize the Identify the correspondence between the image and the abnormal scene; when the abnormal model outputs the abnormal scene corresponding to the image to be identified, confirm that the business scene is abnormal and the abnormal model includes abnormal models corresponding to different business scenarios. For example, when the business scenario is an intelligent transportation business scenario, the abnormal model corresponding to the intelligent transportation business scenario includes a traffic accident model, a traffic congestion model, and an illegal scale type, etc.; when the business scenario is a smart park business scenario, The abnormal model corresponding to the business scene of the smart park includes a personal belongings model, a personnel intrusion model, etc.; when the business scene is a ferry monitoring business scene, the abnormal model corresponding to the ferry monitoring business scene includes an overload model, a falling water model, Illegal ship model, etc.
举例而言,当所述视频图像中当前视频帧出现交通拥堵时,提取所述当前视频帧作为异常图像,并将所述异常图像作为待识别图像导入预先训练好的交通拥堵模型中,当所述交通拥堵模型输出与所述待识别图像对应的交通拥堵场景时,确认所述视频图像对应的智能交通业务场景中出现异常;当所述交通拥堵模型没有输出与所述待识别图像对应的交通拥堵场景时,确认所述视频图像对应的智能交通业务场景中正常。For example, when traffic congestion occurs in the current video frame in the video image, the current video frame is extracted as an abnormal image, and the abnormal image is imported into a pre-trained traffic congestion model as an image to be recognized. When the traffic jam model outputs the traffic jam scene corresponding to the image to be recognized, it is confirmed that there is an abnormality in the intelligent transportation business scene corresponding to the video image; when the traffic jam model does not output the traffic corresponding to the image to be recognized In a congested scene, confirm that the intelligent transportation service scene corresponding to the video image is normal.
上述异常模型为根据图片样本集训练的机器学习模型。所述图片样本包括异常业务场景图片样本和正常业务场景图片样本。所述机器学习模型为可以进行图像识别的人工智能算法模型,包括:卷积神经网络模型CNN、循环神经网络模块RNN和深度神经网络模型DNN。其中,卷积神经网络模型CNN是一种多层神经网络,可以将数据量庞大的图像识别问题不断降维,最终使其能够被训练,因此,本申请实施例中的机器学习模型可以为CNN模型。The above-mentioned abnormal model is a machine learning model trained based on a picture sample set. The picture samples include abnormal business scene picture samples and normal business scene picture samples. The machine learning model is an artificial intelligence algorithm model that can perform image recognition, including: a convolutional neural network model CNN, a recurrent neural network module RNN, and a deep neural network model DNN. Among them, the convolutional neural network model CNN is a multi-layer neural network, which can continuously reduce the dimensionality of the image recognition problem with a huge amount of data, and finally enable it to be trained. Therefore, the machine learning model in the embodiment of the present application may be CNN model.
在CNN网络结构的演化上,出现了许多CNN网络,包括LeNet、AlexNet、VGGNet、GoogleNet和ResNet。其中,ResNet网络提出了一种减轻网络训练负担的残差学习框架,这种网络比以前使用过的网络本质上层次更深,解决了其他神经网络随着网络加深,准确率下降的问题。在本实施方式中,所述机器学习模型可以是卷积申请网络模型CNN中的ResNet模型。需要说明的是,此处仅是举例说明,其他可以进行图像识别的机器学习模型同样适用于本申请,此处不进行赘述。In the evolution of CNN network structure, many CNN networks have appeared, including LeNet, AlexNet, VGGNet, GoogleNet and ResNet. Among them, the ResNet network proposes a residual learning framework that reduces the burden of network training. This network is inherently deeper than the previously used network, and solves the problem of other neural networks that decrease in accuracy as the network deepens. In this embodiment, the machine learning model may be the ResNet model in the convolution application network model CNN. It should be noted that this is only an example, and other machine learning models that can perform image recognition are also applicable to this application, and will not be repeated here.
所述处理模块206用于当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。The processing module 206 is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
在本实施方式中,所述关键信息包括所述业务场景出现异常的时间、地点、及截取的所述视频图像中所述业务场景出现异常时的图片文件等。In this embodiment, the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
进一步地,所述视频分析装置20还可以将记录的关键信息发送至第三方业务平台。所述第三方业务平台包括公安系统、交通管制系统等。通过将所述记录的关键信息发送至所述第三方业务平台,可以帮助所述第三方业务平台及时获取业务场景中出现异常时的关键信息,从而及时处理所述异常。Further, the video analysis device 20 can also send the recorded key information to a third-party service platform. The third-party service platform includes a public security system, a traffic control system, etc. By sending the recorded key information to the third-party business platform, the third-party business platform can help the third-party business platform to obtain key information when an exception occurs in a business scenario in time, so as to process the exception in time.
进一步地,所述所述视频分析装置20还可以展示所述业务场景出现异常时的关键信息。具体地,在显示屏中显示所述业务场景出现异常的时间、地 点、及截取的所述视频图像中所述业务场景出现异常时的图片文件等。Further, the video analysis device 20 can also display key information when the business scene is abnormal. Specifically, the display screen displays the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
综上所述,本申请提供的视频分析装置20,包括接收模块201、检测模块202、跟踪模块203、分析模块204、判断模块205及处理模块206。所述接收模块201用于接收摄像头采集的视频图像;所述检测模块202用于检测所述视频图像中的目标对象得到所述目标对象的类别;所述跟踪模块203用于跟踪所述视频图像中的目标对象得到所述目标对象的状态;所述分析模块204用于根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;所述判断模块205用于判断所述业务场景是否出现异常;及所述处理模块206用于当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。可以实时分析所述视频图像对应的业务场景是否出现异常,并在确认所述业务场景出现异常时,记录所述业务场景出现异常时的关键信息。从而可以将所述关键信息发送至对应的第三方平台,以及时处理所述异常。In summary, the video analysis device 20 provided by the present application includes a receiving module 201, a detection module 202, a tracking module 203, an analysis module 204, a judgment module 205, and a processing module 206. The receiving module 201 is used to receive a video image collected by a camera; the detection module 202 is used to detect a target object in the video image to obtain the target object category; the tracking module 203 is used to track the video image The target object in obtains the status of the target object; the analysis module 204 is configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the status of the target object; the judgment module 205 It is used to determine whether the business scene is abnormal; and the processing module 206 is used to record key information when the business scene is abnormal when the business scene in the video image is abnormal. It is possible to analyze in real time whether the business scene corresponding to the video image is abnormal, and when it is confirmed that the business scene is abnormal, record the key information when the business scene is abnormal. Thereby, the key information can be sent to the corresponding third-party platform, and the exception can be handled in time.
上述以软件功能模块的形式实现的集成的单元,可以存储在一个非易失性可读存储介质中。上述软件功能模块存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,双屏设备,或者网络设备等)或处理器(processor)执行本申请各个实施例所述方法的部分。The aforementioned integrated unit implemented in the form of a software function module may be stored in a non-volatile readable storage medium. The above-mentioned software function module is stored in a storage medium, and includes several instructions to make a computer device (which can be a personal computer, a dual-screen device, or a network device, etc.) or a processor to execute the various embodiments of this application Method part.
实施例三Example three
图3为本申请实施例三提供的计算机设备的示意图。FIG. 3 is a schematic diagram of a computer device provided in Embodiment 3 of this application.
所述计算机设备3包括:数据库31、存储器32、至少一个处理器33、存储在所述存储器32中并可在所述至少一个处理器33上运行的计算机可读指令34及至少一条通讯总线35。The computer device 3 includes: a database 31, a memory 32, at least one processor 33, computer readable instructions 34 stored in the memory 32 and executable on the at least one processor 33, and at least one communication bus 35 .
所述至少一个处理器33执行所述计算机可读指令34时实现上述视频分析方法实施例中的步骤。When the at least one processor 33 executes the computer-readable instructions 34, the steps in the foregoing video analysis method embodiment are implemented.
示例性的,所述计算机可读指令34可以被分割成一个或多个模块/单元,所述一个或者多个模块/单元被存储在所述存储器32中,并由所述至少一个处理器33执行,以完成本申请。所述一个或多个模块/单元可以是能够完成特定功能的一系列计算机可读指令段,该指令段用于描述所述计算机可读指令34在所述计算机设备3中的执行过程。Exemplarily, the computer-readable instructions 34 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 32 and executed by the at least one processor 33 Execute to complete this application. The one or more modules/units may be a series of computer-readable instruction segments capable of completing specific functions, and the instruction segments are used to describe the execution process of the computer-readable instructions 34 in the computer device 3.
所述计算机设备3是一种能够按照事先设定或存储的指令,自动进行数值计算和/或信息处理的设备,其硬件包括但不限于微处理器、专用集成电路(应用程序lication Specific Integrated Circuit,ASIC)、可编程门阵列(Field-Programmable Gate Array,FPGA)、数字处理器(Digital Signal Processor,DSP)、嵌入式设备等。本领域技术人员可以理解,所述示意图3仅仅是计算机设备3的示例,并不构成对计算机设备3的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述计算机设备3还可以包括输入输出设备、网络接入设备、总线等。The computer device 3 is a device that can automatically perform numerical calculation and/or information processing according to pre-set or stored instructions. Its hardware includes, but is not limited to, a microprocessor, an application specific integrated circuit (application license Specific Integrated Circuit). , ASIC), programmable gate array (Field-Programmable Gate Array, FPGA), digital processor (Digital Signal Processor, DSP), embedded equipment, etc. Those skilled in the art can understand that the schematic diagram 3 is only an example of the computer device 3, and does not constitute a limitation on the computer device 3. It may include more or less components than those shown in the figure, or combine certain components, or different components. For example, the computer device 3 may also include input and output devices, network access devices, buses, etc.
所述数据库(Database)31是按照数据结构来组织、存储和管理数据的建立在所述计算机设备3上的仓库。数据库通常分为层次式数据库、网络式 数据库和关系式数据库三种。在本实施方式中,所述数据库31用于存储所述视频图像等。The database (Database) 31 is a warehouse built on the computer device 3 to organize, store and manage data according to a data structure. Databases are usually divided into three types: hierarchical database, network database and relational database. In this embodiment, the database 31 is used to store the video images and the like.
所述至少一个处理器33可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。该处理器33可以是微处理器或者该处理器33也可以是任何常规的处理器等,所述处理器33是所述计算机设备3的控制中心,利用各种接口和线路连接整个计算机设备3的各个部分。The at least one processor 33 may be a central processing unit (Central Processing Unit, CPU), or may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSPs), application specific integrated circuits (ASICs). ), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. The processor 33 can be a microprocessor or the processor 33 can also be any conventional processor, etc. The processor 33 is the control center of the computer device 3, and connects the entire computer device 3 with various interfaces and lines. Parts.
所述存储器32可用于存储所述计算机可读指令34和/或模块/单元,所述处理器33通过运行或执行存储在所述存储器32内的计算机可读指令和/或模块/单元,以及调用存储在存储器32内的数据,实现所述计算机设备3的各种功能。所述存储器32可主要包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序(比如声音播放功能、图像播放功能等)等;存储数据区可存储根据计算机设备3的使用所创建的数据(比如音频数据等)等。此外,存储器32还可以包括非易失性存储器,例如硬盘、内存、插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)、至少一个磁盘存储器件、闪存器件、或其他非易失性固态存储器件。The memory 32 can be used to store the computer-readable instructions 34 and/or modules/units, and the processor 33 runs or executes the computer-readable instructions and/or modules/units stored in the memory 32, and The data stored in the memory 32 is called to realize various functions of the computer device 3. The memory 32 may mainly include a program storage area and a data storage area. The program storage area may store an operating system, an application program required by at least one function (such as a sound playback function, an image playback function, etc.), etc.; The data (such as audio data, etc.) created according to the use of the computer device 3 and the like are stored. In addition, the memory 32 may also include non-volatile memory, such as a hard disk, a memory, a plug-in hard disk, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash Card). , At least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device.
所述存储器32中存储有计算机可读指令代码,且所述至少一个处理器33可调用所述存储器32中存储的计算机可读指令代码以执行相关的功能。例如,图2中所述的各个模块(接收模块201、检测模块202、跟踪模块203、分析模块204、判断模块205及处理模块206)是存储在所述存储器32中的计算机可读指令代码,并由所述至少一个处理器33所执行,从而实现所述各个模块的功能以达到视频分析目的。Computer readable instruction codes are stored in the memory 32, and the at least one processor 33 can call the computer readable instruction codes stored in the memory 32 to perform related functions. For example, the various modules (receiving module 201, detection module 202, tracking module 203, analysis module 204, judgment module 205, and processing module 206) described in FIG. 2 are computer-readable instruction codes stored in the memory 32, It is executed by the at least one processor 33 to realize the functions of the various modules to achieve the purpose of video analysis.
所述接收模块201用于接收摄像头采集的视频图像;The receiving module 201 is used to receive video images collected by a camera;
所述检测模块202用于检测所述视频图像中的目标对象得到所述目标对象的类别;The detection module 202 is configured to detect a target object in the video image to obtain the target object category;
所述跟踪模块203用于跟踪所述视频图像中的目标对象得到所述目标对象的状态;The tracking module 203 is configured to track the target object in the video image to obtain the state of the target object;
所述分析模块204用于根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;The analysis module 204 is configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the state of the target object;
所述判断模块205用于判断所述业务场景是否出现异常;及The judgment module 205 is used to judge whether the business scenario is abnormal; and
所述处理模块206用于当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。The processing module 206 is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
所述计算机设备3集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个非易失性可读取存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,也可以 通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一非易失性可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机可读指令代码,所述计算机可读指令代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述非易失性可读介质可以包括:能够携带所述计算机可读指令代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)等。If the integrated module/unit of the computer device 3 is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a non-volatile readable storage medium. Based on this understanding, this application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program. The computer program can be stored in a non-volatile readable storage medium. When the computer program is executed by the processor, it can implement the steps of the foregoing method embodiments. Wherein, the computer program includes computer readable instruction code, and the computer readable instruction code may be in the form of source code, object code, executable file, or some intermediate form. The non-volatile readable medium may include: any entity or device capable of carrying the computer readable instruction code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) etc.
尽管未示出,所述计算机设备3还可以包括给各个部件供电的电源(比如电池),优选的,电源可以通过电源管理系统与所述至少一个处理器33逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。电源还可以包括一个或一个以上的直流或交流电源、再充电系统、电源故障检测电路、电源转换器或者逆变器、电源状态指示器等任意组件。所述计算机设备3还可以包括蓝牙模块、Wi-Fi模块等,在此不再赘述。Although not shown, the computer device 3 may also include a power source (such as a battery) for supplying power to various components. Preferably, the power source may be logically connected to the at least one processor 33 through a power management system, so as to be implemented through a power management system. Manage functions such as charging, discharging, and power management. The power supply may also include one or more DC or AC power supplies, recharging systems, power failure detection circuits, power converters or inverters, power supply status indicators and other arbitrary components. The computer device 3 may also include a Bluetooth module, a Wi-Fi module, etc., which will not be repeated here.
应该了解,所述实施例仅为说明之用,在专利申请范围上并不受此结构的限制。It should be understood that the described embodiments are for illustrative purposes only, and are not limited by this structure in the scope of the patent application.
在本申请所提供的几个实施例中,应该理解到,所揭露的电子设备和方法,可以通过其它的方式实现。例如,以上所描述的电子设备实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式。In the several embodiments provided in this application, it should be understood that the disclosed electronic device and method may be implemented in other ways. For example, the electronic device embodiments described above are only illustrative. For example, the division of the units is only a logical function division, and there may be other division methods in actual implementation.
另外,在本申请各个实施例中的各功能单元可以集成在相同处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在相同单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能模块的形式实现。In addition, the functional units in the various embodiments of the present application may be integrated in the same processing unit, or each unit may exist alone physically, or two or more units may be integrated in the same unit. The above-mentioned integrated unit can be implemented in the form of hardware or in the form of hardware plus software functional modules.
对于本领域技术人员而言,显然本申请不限于上述示范性实施例的细节,而且在不背离本申请的精神或基本特征的情况下,能够以其他的具体形式实现本申请。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本申请的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化涵括在本申请内。不应将权利要求中的任何附图标记视为限制所涉及的权利要求。此外,显然“包括”一词不排除其他单元或,单数不排除复数。系统权利要求中陈述的多个单元或装置也可以由一个单元或装置通过软件或者硬件来实现。第一,第二等词语用来表示名称,而并不表示任何特定的顺序。For those skilled in the art, it is obvious that the present application is not limited to the details of the foregoing exemplary embodiments, and the present application can be implemented in other specific forms without departing from the spirit or basic characteristics of the application. Therefore, no matter from which point of view, the embodiments should be regarded as exemplary and non-limiting. The scope of this application is defined by the appended claims rather than the above description, and therefore it is intended to fall into the claims. All changes in the meaning and scope of the equivalent elements of are included in this application. Any reference signs in the claims should not be regarded as limiting the claims involved. In addition, it is obvious that the word "including" does not exclude other elements or, and the singular does not exclude the plural. Multiple units or devices stated in the system claims can also be implemented by one unit or device through software or hardware. Words such as first and second are used to denote names, but do not denote any specific order.
最后应说明的是,以上实施例仅用以说明本申请的技术方案而非限制,尽管参照较佳实施例对本申请进行了详细说明,本领域的普通技术人员应当理解,可以对本申请的技术方案进行修改或等同替换,而不脱离本申请技术方案的精神范围。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the application and not to limit them. Although the application has been described in detail with reference to the preferred embodiments, those of ordinary skill in the art should understand that the technical solutions of the application can be Modifications or equivalent replacements are made without departing from the spirit of the technical solution of the present application.

Claims (20)

  1. 一种视频分析方法,其特征在于,所述方法包括:A video analysis method, characterized in that the method includes:
    接收摄像头采集的视频图像;Receive video images collected by the camera;
    检测所述视频图像中的目标对象得到所述目标对象的类别;Detecting the target object in the video image to obtain the target object category;
    跟踪所述视频图像中的目标对象得到所述目标对象的状态;Tracking the target object in the video image to obtain the state of the target object;
    根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;Analyzing the business scene contained in the video image according to the category of the target object and the state of the target object;
    判断所述业务场景是否出现异常;及Determine whether the business scenario is abnormal; and
    当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。When the business scene in the video image is abnormal, the key information when the business scene is abnormal is recorded.
  2. 如权利要求1所述的视频分析方法,其特征在于,所述检测所述视频图像中的目标对象得到所述目标对象的类别包括:The video analysis method of claim 1, wherein the detecting the target object in the video image to obtain the target object category comprises:
    通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性;Obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image;
    将获取的所述基本属性与预先存储在数据库中的目标对象的基本属性进行比对;Comparing the acquired basic attributes with the basic attributes of the target object pre-stored in the database;
    当获取的所述基本属性与所述数据库中的目标对象的基本属性一致时,查询数据库中存储的基本属性与目标对象类别对应表以得到所述目标对象的类别。When the acquired basic attribute is consistent with the basic attribute of the target object in the database, query a table corresponding to the basic attribute and target object category stored in the database to obtain the target object category.
  3. 如权利要求1所述的视频分析方法,其特征在于,所述跟踪所述视频图像中的目标对象得到所述目标对象的状态包括:The video analysis method of claim 1, wherein the tracking the target object in the video image to obtain the state of the target object comprises:
    确定当前视频帧中的目标对象;Determine the target object in the current video frame;
    获取目标对象在前序视频帧中的图像区域以及所述图像区域的图像特征,其中,所述前序视频帧为当前视频帧之前的k个视频帧,k为正整数;Acquiring the image area of the target object in the previous video frame and the image characteristics of the image area, where the previous video frame is k video frames before the current video frame, and k is a positive integer;
    根据所述目标对象在前序视频帧中的图像区域,对所述目标对象进行运动估计,确定所述目标对象在当前视频帧的预测区域;Perform motion estimation on the target object according to the image area of the target object in the previous video frame, and determine the prediction area of the target object in the current video frame;
    根据所述预测区域确定目标对象在当前视频帧中的检测范围;Determine the detection range of the target object in the current video frame according to the prediction area;
    判断所述目标对象是否出现在当前视频帧中的检测范围;Judging whether the target object appears in the detection range of the current video frame;
    若所述目标对象出现在当前视频帧中的检测范围,确定所述目标对象在当前视频帧中的图像区域;If the target object appears in the detection range of the current video frame, determine the image area of the target object in the current video frame;
    若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象异常。If the target object does not appear in the detection range in the current video frame, it is determined that the target object is abnormal.
  4. 如权利要求3所述的视频分析方法,其特征在于,所述判断所述业务场景是否出现异常包括:5. The video analysis method according to claim 3, wherein said judging whether the business scene is abnormal comprises:
    当确定所述目标对象异常时,提取所述当前视频帧作为异常图像;When it is determined that the target object is abnormal, extract the current video frame as an abnormal image;
    将所述异常图像作为待识别图像导入预先训练好的异常模型中,其中,所述异常模型用于表征待识别图像与异常场景之间的对应关系;Importing the abnormal image as an image to be identified into a pre-trained anomaly model, where the abnormal model is used to characterize the correspondence between the image to be identified and the abnormal scene;
    当所述异常模型输出与所述待识别图像对应的异常场景时,确认所述业务场景出现异常。When the abnormal model outputs the abnormal scene corresponding to the image to be recognized, it is confirmed that the business scene is abnormal.
  5. 如权利要求1所述的视频分析方法,其特征在于,所述关键信息包括所述业务场景出现异常的时间、地点、及截取的所述视频图像中所述业务场景出现异常时的图片文件。The video analysis method according to claim 1, wherein the key information includes the time and place when the business scene is abnormal, and the picture file when the business scene is abnormal in the intercepted video image.
  6. 如权利要求5所述的视频分析方法,其特征在于,所述方法还包括:5. The video analysis method of claim 5, wherein the method further comprises:
    发送记录的关键信息至第三方业务平台,其中,所述第三方业务平台包括公安系统和交通管制系统。Send the recorded key information to a third-party business platform, where the third-party business platform includes a public security system and a traffic control system.
  7. 如权利要求1所述的视频分析方法,其特征在于,在接收摄像头采集的视频图像后,所述方法还包括:The video analysis method of claim 1, wherein after receiving the video image collected by the camera, the method further comprises:
    对所述视频图像进行解码。Decoding the video image.
  8. 一种视频分析装置,其特征在于,所述装置包括:A video analysis device, characterized in that the device includes:
    接收模块,用于接收摄像头采集的视频图像;The receiving module is used to receive the video image collected by the camera;
    检测模块,用于检测所述视频图像中的目标对象得到所述目标对象的类别;The detection module is used to detect the target object in the video image to obtain the target object category;
    跟踪模块,用于跟踪所述视频图像中的目标对象得到所述目标对象的状态;A tracking module, used to track the target object in the video image to obtain the state of the target object;
    分析模块,用于根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;An analysis module, configured to analyze and obtain the business scene contained in the video image according to the category of the target object and the state of the target object;
    判断模块,用于判断所述业务场景是否出现异常;及The judgment module is used to judge whether the business scenario is abnormal; and
    处理模块,用于当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。The processing module is configured to record key information when the business scene is abnormal when the business scene in the video image is abnormal.
  9. 一种计算机设备,其特征在于,所述计算机设备包括处理器和存储器,所述存储器中存储至少一个计算机可读指令,所述处理器执行所述至少一个计算机可读指令以实现以下步骤:A computer device, characterized in that the computer device includes a processor and a memory, the memory stores at least one computer readable instruction, and the processor executes the at least one computer readable instruction to implement the following steps:
    接收摄像头采集的视频图像;Receive video images collected by the camera;
    检测所述视频图像中的目标对象得到所述目标对象的类别;Detecting the target object in the video image to obtain the target object category;
    跟踪所述视频图像中的目标对象得到所述目标对象的状态;Tracking the target object in the video image to obtain the state of the target object;
    根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;Analyzing the business scene contained in the video image according to the category of the target object and the state of the target object;
    判断所述业务场景是否出现异常;及Determine whether the business scenario is abnormal; and
    当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。When the business scene in the video image is abnormal, the key information when the business scene is abnormal is recorded.
  10. 如权利要求9所述的计算机设备,其特征在于,所述处理器执行所述至少一个计算机可读指令以实现所述检测所述视频图像中的目标对象得到所述目标对象的类别时,包括:The computer device according to claim 9, wherein when the processor executes the at least one computer-readable instruction to realize the detection of the target object in the video image to obtain the target object category, the method comprises :
    通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性;Obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image;
    将获取的所述基本属性与预先存储在数据库中的目标对象的基本属性进行比对;Comparing the acquired basic attributes with the basic attributes of the target object pre-stored in the database;
    当获取的所述基本属性与所述数据库中的目标对象的基本属性一致时,查 询数据库中存储的基本属性与目标对象类别对应表以得到所述目标对象的类别。When the obtained basic attribute is consistent with the basic attribute of the target object in the database, the corresponding table of the basic attribute and the target object category stored in the database is queried to obtain the target object category.
  11. 如权利要求9所述的计算机设备,其特征在于,所述处理器执行所述至少一个计算机可读指令以实现所述跟踪所述视频图像中的目标对象得到所述目标对象的状态时,包括:The computer device according to claim 9, wherein when the processor executes the at least one computer-readable instruction to realize the tracking of the target object in the video image to obtain the state of the target object, it comprises :
    确定当前视频帧中的目标对象;Determine the target object in the current video frame;
    获取目标对象在前序视频帧中的图像区域以及所述图像区域的图像特征,其中,所述前序视频帧为当前视频帧之前的k个视频帧,k为正整数;Acquiring the image area of the target object in the previous video frame and the image characteristics of the image area, where the previous video frame is k video frames before the current video frame, and k is a positive integer;
    根据所述目标对象在前序视频帧中的图像区域,对所述目标对象进行运动估计,确定所述目标对象在当前视频帧的预测区域;Perform motion estimation on the target object according to the image area of the target object in the previous video frame, and determine the prediction area of the target object in the current video frame;
    根据所述预测区域确定目标对象在当前视频帧中的检测范围;Determine the detection range of the target object in the current video frame according to the prediction area;
    判断所述目标对象是否出现在当前视频帧中的检测范围;Judging whether the target object appears in the detection range of the current video frame;
    若所述目标对象出现在当前视频帧中的检测范围,确定所述目标对象在当前视频帧中的图像区域;If the target object appears in the detection range of the current video frame, determine the image area of the target object in the current video frame;
    若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象异常。If the target object does not appear in the detection range in the current video frame, it is determined that the target object is abnormal.
  12. 如权利要求11所述的计算机设备,其特征在于,所述处理器执行所述至少一个计算机可读指令以实现所述判断所述业务场景是否出现异常时,包括:The computer device according to claim 11, wherein when the processor executes the at least one computer-readable instruction to realize the judging whether the business scenario is abnormal, it comprises:
    当确定所述目标对象异常时,提取所述当前视频帧作为异常图像;When it is determined that the target object is abnormal, extract the current video frame as an abnormal image;
    将所述异常图像作为待识别图像导入预先训练好的异常模型中,其中,所述异常模型用于表征待识别图像与异常场景之间的对应关系;Importing the abnormal image as an image to be identified into a pre-trained anomaly model, where the abnormal model is used to characterize the correspondence between the image to be identified and the abnormal scene;
    当所述异常模型输出与所述待识别图像对应的异常场景时,确认所述业务场景出现异常。When the abnormal model outputs the abnormal scene corresponding to the image to be recognized, it is confirmed that the business scene is abnormal.
  13. 如权利要求9所述的计算机设备,其特征在于,所述处理器执行所述至少一个计算机可读指令时还用以实现以下步骤:9. The computer device of claim 9, wherein the processor is further configured to implement the following steps when executing the at least one computer readable instruction:
    发送记录的关键信息至第三方业务平台,其中,所述第三方业务平台包括公安系统和交通管制系统。Send the recorded key information to a third-party business platform, where the third-party business platform includes a public security system and a traffic control system.
  14. 如权利要求9所述的计算机设备,其特征在于,所述处理器执行所述至少一个计算机可读指令以实现在所述接收摄像头采集的视频图像后,还用以实现以下步骤:9. The computer device according to claim 9, wherein the processor executes the at least one computer-readable instruction to implement the following steps after receiving the video image collected by the camera:
    对所述视频图像进行解码。Decoding the video image.
  15. 一种非易失性可读存储介质,所述非易失性可读存储介质上存储有至少一个计算机可读指令,其特征在于,所述至少一个计算机可读指令被处理器执行时以实现以下步骤:A non-volatile readable storage medium having at least one computer readable instruction stored on the non-volatile readable storage medium, wherein the at least one computer readable instruction is executed by a processor to realize The following steps:
    接收摄像头采集的视频图像;Receive video images collected by the camera;
    检测所述视频图像中的目标对象得到所述目标对象的类别;Detecting the target object in the video image to obtain the target object category;
    跟踪所述视频图像中的目标对象得到所述目标对象的状态;Tracking the target object in the video image to obtain the state of the target object;
    根据所述目标对象的类别和所述目标对象的状态分析得到所述视频图像中包含的业务场景;Analyzing the business scene contained in the video image according to the category of the target object and the state of the target object;
    判断所述业务场景是否出现异常;及Determine whether the business scenario is abnormal; and
    当所述视频图像中的业务场景出现异常时,记录所述业务场景出现异常时的关键信息。When the business scene in the video image is abnormal, the key information when the business scene is abnormal is recorded.
  16. 如权利要求15所述的存储介质,其特征在于,所述至少一个计算机可读指令被处理器执行以实现所述检测所述视频图像中的目标对象得到所述目标对象的类别时,包括:15. The storage medium of claim 15, wherein when the at least one computer-readable instruction is executed by a processor to realize the detection of the target object in the video image to obtain the target object category, the method comprises:
    通过分解所述视频图像中的目标对象,获取所述视频图像中的目标对象的基本属性;Obtaining the basic attributes of the target object in the video image by decomposing the target object in the video image;
    将获取的所述基本属性与预先存储在数据库中的目标对象的基本属性进行比对;Comparing the acquired basic attributes with the basic attributes of the target object pre-stored in the database;
    当获取的所述基本属性与所述数据库中的目标对象的基本属性一致时,查询数据库中存储的基本属性与目标对象类别对应表以得到所述目标对象的类别。When the acquired basic attribute is consistent with the basic attribute of the target object in the database, query a table corresponding to the basic attribute and target object category stored in the database to obtain the target object category.
  17. 如权利要求15所述的存储介质,其特征在于,所述至少一个计算机可读指令被处理器执行以实现所述跟踪所述视频图像中的目标对象得到所述目标对象的状态时,包括:15. The storage medium of claim 15, wherein when the at least one computer-readable instruction is executed by a processor to implement the tracking of the target object in the video image to obtain the state of the target object, the method comprises:
    确定当前视频帧中的目标对象;Determine the target object in the current video frame;
    获取目标对象在前序视频帧中的图像区域以及所述图像区域的图像特征,其中,所述前序视频帧为当前视频帧之前的k个视频帧,k为正整数;Acquiring the image area of the target object in the previous video frame and the image characteristics of the image area, where the previous video frame is k video frames before the current video frame, and k is a positive integer;
    根据所述目标对象在前序视频帧中的图像区域,对所述目标对象进行运动估计,确定所述目标对象在当前视频帧的预测区域;Perform motion estimation on the target object according to the image area of the target object in the previous video frame, and determine the prediction area of the target object in the current video frame;
    根据所述预测区域确定目标对象在当前视频帧中的检测范围;Determine the detection range of the target object in the current video frame according to the prediction area;
    判断所述目标对象是否出现在当前视频帧中的检测范围;Judging whether the target object appears in the detection range of the current video frame;
    若所述目标对象出现在当前视频帧中的检测范围,确定所述目标对象在当前视频帧中的图像区域;If the target object appears in the detection range of the current video frame, determine the image area of the target object in the current video frame;
    若所述目标对象没有出现在当前视频帧中的检测范围,确定所述目标对象异常。If the target object does not appear in the detection range in the current video frame, it is determined that the target object is abnormal.
  18. 如权利要求17所述的存储介质,其特征在于,所述至少一个计算机可读指令被处理器执行以实现所述判断所述业务场景是否出现异常时,包括:The storage medium according to claim 17, wherein when the at least one computer-readable instruction is executed by the processor to realize the judging whether the business scenario is abnormal, it comprises:
    当确定所述目标对象异常时,提取所述当前视频帧作为异常图像;When it is determined that the target object is abnormal, extract the current video frame as an abnormal image;
    将所述异常图像作为待识别图像导入预先训练好的异常模型中,其中,所述异常模型用于表征待识别图像与异常场景之间的对应关系;Importing the abnormal image as an image to be identified into a pre-trained anomaly model, where the abnormal model is used to characterize the correspondence between the image to be identified and the abnormal scene;
    当所述异常模型输出与所述待识别图像对应的异常场景时,确认所述业务场景出现异常。When the abnormal model outputs the abnormal scene corresponding to the image to be recognized, it is confirmed that the business scene is abnormal.
  19. 如权利要求15所述的存储介质,其特征在于,所述至少一个计算机可读指令被处理器执行时还用以实现以下步骤:15. The storage medium of claim 15, wherein the at least one computer readable instruction is further used to implement the following steps when executed by the processor:
    发送记录的关键信息至第三方业务平台,其中,所述第三方业务平台包括公安系统和交通管制系统。Send the recorded key information to a third-party business platform, where the third-party business platform includes a public security system and a traffic control system.
  20. 如权利要求15所述的存储介质,其特征在于,所述至少一个计算机可读指令被处理器执行以实现在所述接收摄像头采集的视频图像后,还用以 实现以下步骤:The storage medium according to claim 15, wherein the at least one computer-readable instruction is executed by the processor to implement the following steps after the video image collected by the camera is received:
    对所述视频图像进行解码。Decoding the video image.
PCT/CN2019/103373 2019-06-14 2019-08-29 Video analysis method and apparatus, computer device and storage medium WO2020248386A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910517477.X 2019-06-14
CN201910517477.XA CN110390262B (en) 2019-06-14 2019-06-14 Video analysis method, device, server and storage medium

Publications (1)

Publication Number Publication Date
WO2020248386A1 true WO2020248386A1 (en) 2020-12-17

Family

ID=68285438

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/103373 WO2020248386A1 (en) 2019-06-14 2019-08-29 Video analysis method and apparatus, computer device and storage medium

Country Status (2)

Country Link
CN (1) CN110390262B (en)
WO (1) WO2020248386A1 (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112633126A (en) * 2020-12-18 2021-04-09 联通物联网有限责任公司 Video processing method and device
CN112634329A (en) * 2020-12-26 2021-04-09 西安电子科技大学 Scene target activity prediction method and device based on space-time and-or graph
CN112711994A (en) * 2020-12-21 2021-04-27 航天信息股份有限公司 Method and system for detecting illegal operation behaviors based on scene recognition
CN112749636A (en) * 2020-12-29 2021-05-04 精英数智科技股份有限公司 Method, device and system for monitoring water exploration and drainage of coal mine and storage medium
CN112991280A (en) * 2021-03-03 2021-06-18 望知科技(深圳)有限公司 Visual detection method and system and electronic equipment
CN113065456A (en) * 2021-03-30 2021-07-02 上海商汤智能科技有限公司 Information prompting method and device, electronic equipment and computer storage medium
CN113361468A (en) * 2021-06-30 2021-09-07 北京百度网讯科技有限公司 Business quality inspection method, device, equipment and storage medium
CN113378005A (en) * 2021-06-03 2021-09-10 北京百度网讯科技有限公司 Event processing method and device, electronic equipment and storage medium
CN113422935A (en) * 2021-07-06 2021-09-21 城云科技(中国)有限公司 Video stream processing method, device and system
CN113705370A (en) * 2021-08-09 2021-11-26 百度在线网络技术(北京)有限公司 Method and device for detecting illegal behavior of live broadcast room, electronic equipment and storage medium
CN113763860A (en) * 2021-09-14 2021-12-07 杭州海康消防科技有限公司 Display color determination method and device, electronic equipment and storage medium
CN113992890A (en) * 2021-10-22 2022-01-28 北京明略昭辉科技有限公司 Monitoring method, monitoring device, storage medium and electronic equipment
CN114205565A (en) * 2022-02-15 2022-03-18 云丁网络技术(北京)有限公司 Monitoring video distribution method and system
CN114378862A (en) * 2022-03-02 2022-04-22 北京云迹科技股份有限公司 Robot abnormity automatic repairing method and device based on cloud platform and robot
CN114682520A (en) * 2022-04-12 2022-07-01 浪潮软件集团有限公司 Substandard product sorting device based on domestic CPU and artificial intelligence accelerator card
WO2023040151A1 (en) * 2021-09-17 2023-03-23 上海商汤智能科技有限公司 Algorithm application element generating method and apparatus, electronic device, computer readable storage medium, and computer program product
CN116953416A (en) * 2023-09-19 2023-10-27 英迪格(天津)电气有限公司 Monitoring system for running state of railway power transformation and distribution device
CN117079079A (en) * 2023-09-27 2023-11-17 中电科新型智慧城市研究院有限公司 Training method of video anomaly detection model, video anomaly detection method and system
CN114378862B (en) * 2022-03-02 2024-05-10 北京云迹科技股份有限公司 Cloud platform-based automatic robot abnormality repairing method and device and robot

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111339894A (en) * 2020-02-20 2020-06-26 支付宝(杭州)信息技术有限公司 Data processing and risk identification method, device, equipment and medium
CN111652043A (en) * 2020-04-15 2020-09-11 北京三快在线科技有限公司 Object state identification method and device, image acquisition equipment and storage medium
CN113552123A (en) * 2020-04-17 2021-10-26 华为技术有限公司 Visual inspection method and visual inspection device
CN111680610A (en) * 2020-06-03 2020-09-18 合肥中科类脑智能技术有限公司 Construction scene abnormity monitoring method and device
CN111783591B (en) * 2020-06-23 2024-04-26 北京百度网讯科技有限公司 Abnormality detection method, abnormality detection device, abnormality detection apparatus, and recording medium
CN111832492A (en) * 2020-07-16 2020-10-27 平安科技(深圳)有限公司 Method and device for distinguishing static traffic abnormality, computer equipment and storage medium
CN112804489B (en) * 2020-12-31 2023-02-17 重庆文理学院 Intelligent construction site management system and method based on Internet +
CN113891072B (en) * 2021-12-08 2022-02-11 北京拙河科技有限公司 Video monitoring and anomaly analysis system and method based on hundred million-level pixel data
CN114792368A (en) * 2022-04-28 2022-07-26 上海兴容信息技术有限公司 Method and system for intelligently judging store compliance
CN116708899B (en) * 2022-06-30 2024-01-23 北京生数科技有限公司 Video processing method, device and storage medium applied to virtual image synthesis
CN115834621A (en) * 2022-11-16 2023-03-21 山东新一代信息产业技术研究院有限公司 Accident quick-place device and method based on artificial intelligence

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107194318A (en) * 2017-04-24 2017-09-22 北京航空航天大学 The scene recognition method of target detection auxiliary
CN107346415A (en) * 2017-06-08 2017-11-14 小草数语(北京)科技有限公司 Method of video image processing, device and monitoring device
CN109063667A (en) * 2018-08-14 2018-12-21 视云融聚(广州)科技有限公司 A kind of video identification method optimizing and method for pushing based on scene
US20190095716A1 (en) * 2017-09-26 2019-03-28 Ambient AI, Inc Systems and methods for intelligent and interpretive analysis of video image data using machine learning
CN109598885A (en) * 2018-12-21 2019-04-09 广东中安金狮科创有限公司 Monitoring system and its alarm method

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8200011B2 (en) * 2007-09-27 2012-06-12 Behavioral Recognition Systems, Inc. Context processor for video analysis system
CN108830204B (en) * 2018-06-01 2021-10-19 中国科学技术大学 Method for detecting abnormality in target-oriented surveillance video

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107194318A (en) * 2017-04-24 2017-09-22 北京航空航天大学 The scene recognition method of target detection auxiliary
CN107346415A (en) * 2017-06-08 2017-11-14 小草数语(北京)科技有限公司 Method of video image processing, device and monitoring device
US20190095716A1 (en) * 2017-09-26 2019-03-28 Ambient AI, Inc Systems and methods for intelligent and interpretive analysis of video image data using machine learning
CN109063667A (en) * 2018-08-14 2018-12-21 视云融聚(广州)科技有限公司 A kind of video identification method optimizing and method for pushing based on scene
CN109598885A (en) * 2018-12-21 2019-04-09 广东中安金狮科创有限公司 Monitoring system and its alarm method

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112633126A (en) * 2020-12-18 2021-04-09 联通物联网有限责任公司 Video processing method and device
CN112711994A (en) * 2020-12-21 2021-04-27 航天信息股份有限公司 Method and system for detecting illegal operation behaviors based on scene recognition
CN112634329A (en) * 2020-12-26 2021-04-09 西安电子科技大学 Scene target activity prediction method and device based on space-time and-or graph
CN112634329B (en) * 2020-12-26 2024-02-13 西安电子科技大学 Scene target activity prediction method and device based on space-time and or graph
CN112749636A (en) * 2020-12-29 2021-05-04 精英数智科技股份有限公司 Method, device and system for monitoring water exploration and drainage of coal mine and storage medium
CN112749636B (en) * 2020-12-29 2023-10-31 精英数智科技股份有限公司 Monitoring method, device and system for water drainage detection of coal mine and storage medium
CN112991280A (en) * 2021-03-03 2021-06-18 望知科技(深圳)有限公司 Visual detection method and system and electronic equipment
CN113065456A (en) * 2021-03-30 2021-07-02 上海商汤智能科技有限公司 Information prompting method and device, electronic equipment and computer storage medium
CN113378005B (en) * 2021-06-03 2023-06-02 北京百度网讯科技有限公司 Event processing method, device, electronic equipment and storage medium
CN113378005A (en) * 2021-06-03 2021-09-10 北京百度网讯科技有限公司 Event processing method and device, electronic equipment and storage medium
CN113361468A (en) * 2021-06-30 2021-09-07 北京百度网讯科技有限公司 Business quality inspection method, device, equipment and storage medium
CN113422935A (en) * 2021-07-06 2021-09-21 城云科技(中国)有限公司 Video stream processing method, device and system
CN113705370B (en) * 2021-08-09 2023-06-30 百度在线网络技术(北京)有限公司 Method and device for detecting illegal behaviors of live broadcasting room, electronic equipment and storage medium
CN113705370A (en) * 2021-08-09 2021-11-26 百度在线网络技术(北京)有限公司 Method and device for detecting illegal behavior of live broadcast room, electronic equipment and storage medium
CN113763860A (en) * 2021-09-14 2021-12-07 杭州海康消防科技有限公司 Display color determination method and device, electronic equipment and storage medium
WO2023040151A1 (en) * 2021-09-17 2023-03-23 上海商汤智能科技有限公司 Algorithm application element generating method and apparatus, electronic device, computer readable storage medium, and computer program product
CN113992890A (en) * 2021-10-22 2022-01-28 北京明略昭辉科技有限公司 Monitoring method, monitoring device, storage medium and electronic equipment
CN114205565B (en) * 2022-02-15 2022-07-29 云丁网络技术(北京)有限公司 Monitoring video distribution method and system
CN114205565A (en) * 2022-02-15 2022-03-18 云丁网络技术(北京)有限公司 Monitoring video distribution method and system
CN114378862A (en) * 2022-03-02 2022-04-22 北京云迹科技股份有限公司 Robot abnormity automatic repairing method and device based on cloud platform and robot
CN114378862B (en) * 2022-03-02 2024-05-10 北京云迹科技股份有限公司 Cloud platform-based automatic robot abnormality repairing method and device and robot
CN114682520A (en) * 2022-04-12 2022-07-01 浪潮软件集团有限公司 Substandard product sorting device based on domestic CPU and artificial intelligence accelerator card
CN116953416A (en) * 2023-09-19 2023-10-27 英迪格(天津)电气有限公司 Monitoring system for running state of railway power transformation and distribution device
CN116953416B (en) * 2023-09-19 2023-12-08 英迪格(天津)电气有限公司 Monitoring system for running state of railway power transformation and distribution device
CN117079079B (en) * 2023-09-27 2024-03-15 中电科新型智慧城市研究院有限公司 Training method of video anomaly detection model, video anomaly detection method and system
CN117079079A (en) * 2023-09-27 2023-11-17 中电科新型智慧城市研究院有限公司 Training method of video anomaly detection model, video anomaly detection method and system

Also Published As

Publication number Publication date
CN110390262B (en) 2023-06-30
CN110390262A (en) 2019-10-29

Similar Documents

Publication Publication Date Title
WO2020248386A1 (en) Video analysis method and apparatus, computer device and storage medium
US11840239B2 (en) Multiple exposure event determination
US10706330B2 (en) Methods and systems for accurately recognizing vehicle license plates
WO2021135879A1 (en) Vehicle data monitoring method and apparatus, computer device, and storage medium
US10552687B2 (en) Visual monitoring of queues using auxillary devices
CN109191829B (en) road safety monitoring method and system, and computer readable storage medium
US20230153698A1 (en) Methods and systems for accurately recognizing vehicle license plates
CN109360362A (en) A kind of railway video monitoring recognition methods, system and computer-readable medium
CN102902960B (en) Leave-behind object detection method based on Gaussian modelling and target contour
US20160210759A1 (en) System and method of detecting moving objects
CN112233428B (en) Traffic flow prediction method, device, storage medium and equipment
CN111079621A (en) Method and device for detecting object, electronic equipment and storage medium
WO2021022698A1 (en) Following detection method and apparatus, and electronic device and storage medium
CN114926791A (en) Method and device for detecting abnormal lane change of vehicles at intersection, storage medium and electronic equipment
CN113380021A (en) Vehicle state detection method, device, server and computer-readable storage medium
CN114663871A (en) Image recognition method, training method, device, system and storage medium
Jiao et al. Traffic behavior recognition from traffic videos under occlusion condition: a Kalman filter approach
CN103761345A (en) Video retrieval method based on OCR character recognition technology
CN114693722B (en) Vehicle driving behavior detection method, detection device and detection equipment
EP4071728A1 (en) Artificial intelligence model integration and deployment for providing a service
CN115019242A (en) Abnormal event detection method and device for traffic scene and processing equipment
CN112153341B (en) Task supervision method, device and system, electronic equipment and storage medium
Huang et al. A bus crowdedness sensing system using deep-learning based object detection
Maaloul Video-based algorithms for accident detections
CN115240406B (en) Road congestion management method and device, computer readable medium and electronic equipment

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19933143

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19933143

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 19933143

Country of ref document: EP

Kind code of ref document: A1