WO2022105517A1 - Systems and methods for detecting traffic accidents - Google Patents
Systems and methods for detecting traffic accidents Download PDFInfo
- Publication number
- WO2022105517A1 WO2022105517A1 PCT/CN2021/124913 CN2021124913W WO2022105517A1 WO 2022105517 A1 WO2022105517 A1 WO 2022105517A1 CN 2021124913 W CN2021124913 W CN 2021124913W WO 2022105517 A1 WO2022105517 A1 WO 2022105517A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- static
- target image
- target
- image
- target scene
- Prior art date
Links
- 206010039203 Road traffic accident Diseases 0.000 title claims abstract description 150
- 238000000034 method Methods 0.000 title claims description 156
- 230000003068 static effect Effects 0.000 claims abstract description 290
- 238000012545 processing Methods 0.000 claims abstract description 188
- 238000001514 detection method Methods 0.000 claims description 99
- 238000003860 storage Methods 0.000 claims description 73
- 230000004044 response Effects 0.000 claims description 31
- 230000011218 segmentation Effects 0.000 claims description 22
- 230000004927 fusion Effects 0.000 claims description 17
- 230000009466 transformation Effects 0.000 claims description 13
- 238000004891 communication Methods 0.000 claims description 9
- 230000008569 process Effects 0.000 description 105
- 238000010586 diagram Methods 0.000 description 28
- 230000015654 memory Effects 0.000 description 11
- 238000012549 training Methods 0.000 description 10
- 238000012986 modification Methods 0.000 description 8
- 230000004048 modification Effects 0.000 description 8
- 230000006870 function Effects 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 5
- 238000009499 grossing Methods 0.000 description 4
- 230000014759 maintenance of location Effects 0.000 description 4
- 238000003062 neural network model Methods 0.000 description 4
- 230000006872 improvement Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 230000004075 alteration Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 206010033799 Paralysis Diseases 0.000 description 1
- 241001464837 Viridiplantae Species 0.000 description 1
- 230000032683 aging Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000001174 ascending effect Effects 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 238000000429 assembly Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 230000006837 decompression Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000010977 jade Substances 0.000 description 1
- 230000007257 malfunction Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
- G06V20/54—Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/49—Segmenting video sequences, i.e. computational techniques such as parsing or cutting the sequence, low-level clustering or determining units such as shots or scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- the present disclosure generally relates to image processing, and in particular, to systems and methods for detecting a traffic accident by processing at least one image.
- a current accident detection system mainly relies on manual observation, which is difficult to simultaneously monitor a large number of traffic scenes and efficiently detect traffic accidents therein. Therefore, it is desirable to provide systems and methods for automatically and efficiently detecting traffic accidents in traffic scenes, thereby improving the efficiency of traffic supervision.
- a system may be provided.
- the system may include: at least one storage device including a set of instructions; and at least one processor in communication with the at least one storage device, wherein when executing the set of instructions, the at least one processor is configured to cause the system to: obtain a video of a target scene that is captured by an image capture device; determine a target image associated with the target scene based on the video; determine a static region associated with the target scene in the target image; and determine a result indicating whether a traffic accident happens in the static region by processing, based on a first model, the static region associated with the target scene in the target image.
- the at least one processor may be configured to cause the system to: obtain, from the video, a first image; obtain, from the video, one or more second images captured prior to the first image; and determine the target image by processing a dynamic region associated with the target scene in the first image based on the one or more second images.
- the at least one processor may be configured to cause the system to: generate a fusion result by fusing the first image and the one or more second images; and determine the target image based on the fusion result.
- a time interval between two adjacent images among the one or more second images and the first image may relate to a type of the target scene.
- the at least one processor may be configured to cause the system to: determine one or more regions associated with one or more static objects of the target scene in the target image by processing the target image based on a second model; and determine the static region associated with the target scene in the target image based on the one or more regions associated with the one or more static objects of the target scene in the target image.
- the second model may include a detection model; and the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more bounding boxes of the one or more static objects determined by the detection model.
- the second model may include a segmentation model; and the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more connected components of the one or more static objects determined by the segmentation model.
- the at least one processor may be configured to cause the system to: determine at least one of the one or more regions that satisfies a first condition; and generate a combination result by combining the at least one region; and determine the static region associated with the target scene in the target image based on the combination result.
- the first condition may relate to at least one distance between the one or more regions.
- the at least one processor may be configured to cause the system to: determine a confidence value associated with the traffic accident by processing, based on the first model, the static region associated with the target scene in the target image; and in response to determining that the confidence value satisfies a second condition, determine that the traffic accident happens in the static region.
- the at least one processor may be configured to cause the system to: in response to determining that the confidence value fails to satisfy the second condition, determine whether there are at least two of the one or more regions associated with the one or more static objects in the static region that overlap; and in response to determining that there are the at least two of the one or more regions associated with the one or more static objects in the static region that overlap, modify the confidence value associated with the traffic accident.
- the at least one processor may be configured to cause the system to: determine one or more three-dimensional regions associated with the one or more static objects in the static region by performing a three-dimensional transformation on the one or more two-dimensional regions associated with the one or more static objects based on one or more coordinates associated with the one or more two-dimensional regions; and modify the confidence value associated with the traffic accident based on an overlap between the one or more three-dimensional regions associated with the one or more static objects in the static region.
- a method may be provided.
- the method may be implemented on a computing device having at least one processor, at least one storage medium, and a communication platform connected to a network.
- the method may include: obtaining a video of a target scene that is captured by an image capture device; determining a target image associated with the target scene based on the video; determining a static region associated with the target scene in the target image; and determining a result indicating whether a traffic accident happens in the static region by processing, based on a first model, the static region associated with the target scene in the target image.
- determining a target image associated with the target scene based on the video may include: obtaining, from the video, a first image; obtaining, from the video, one or more second images captured prior to the first image; and determining the target image by processing a dynamic region associated with the target scene in the first image based on the one or more second images.
- determining the target image by processing a dynamic region associated with the target scene in the first image based on the one or more second images may include: generating a fusion result by fusing the first image and the one or more second images; and determining the target image based on the fusion result.
- a time interval between two adjacent images among the one or more second images and the first image may relate to a type of the target scene.
- determining a static region associated with the target scene in the target image may include: determining one or more regions associated with one or more static objects of the target scene in the target image by processing the target image based on a second model; and determining the static region associated with the target scene in the target image based on the one or more regions associated with the one or more static objects of the target scene in the target image.
- the second model may include a detection model; and the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more bounding boxes of the one or more static objects determined by the detection model.
- the second model may include a segmentation model; and the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more connected components of the one or more static objects determined by the segmentation model.
- determining the static region associated with the target scene in the target image based on the one or more regions associated with the one or more static objects of the target scene in the target image may include: determining at least one of the one or more regions that satisfies a first condition; and generating a combination result by combining the at least one region; and determining the static region associated with the target scene in the target image based on the combination result.
- the first condition may relate to at least one distance between the one or more regions.
- the static region associated with the target scene in the target image may include: determining a confidence value associated with the traffic accident by processing, based on the first model, the static region associated with the target scene in the target image; and in response to determining that the confidence value satisfies a second condition, determining that the traffic accident happens in the static region.
- the static region associated with the target scene in the target image may include: in response to determining that the confidence value fails to satisfy the second condition, determining whether there are at least two of the one or more regions associated with the one or more static objects in the static region that overlap; and in response to determining that there are the at least two of the one or more regions associated with the one or more static objects in the static region that overlap, modifying the confidence value associated with the traffic accident.
- the one or more regions associated with the one or more static objects may include one or more two-dimensional regions; and the modifying the confidence value associated with the traffic accident may include: determining one or more three-dimensional regions associated with the one or more static objects in the static region by performing a three-dimensional transformation on the one or more two-dimensional regions associated with the one or more static objects based on one or more coordinates associated with the one or more two-dimensional regions; and modifying the confidence value associated with the traffic accident based on an overlap between the one or more three-dimensional regions associated with the one or more static objects in the static region.
- a non-transitory computer readable medium may include executable instructions that, when executed by at least one processor, directs the at least one processor to perform a method.
- the method may include: obtaining a video of a target scene that is captured by an image capture device; determining a target image associated with the target scene based on the video; determining a static region associated with the target scene in the target image; and determining a result indicating whether a traffic accident happens in the static region by processing, based on a first model, the static region associated with the target scene in the target image.
- FIG. 1 is a schematic diagram illustrating an exemplary accident detection system according to some embodiments of the present disclosure
- FIG. 2 is a schematic diagram illustrating exemplary hardware and/or software components of an exemplary computing device according to some embodiments of the present disclosure
- FIG. 3 is a schematic diagram illustrating exemplary hardware and/or software components of an exemplary mobile device according to some embodiments of the present disclosure
- FIG. 4 is a block diagram illustrating an exemplary processing device according to some embodiments of the present disclosure.
- FIG. 5 is a flowchart illustrating an exemplary process for determining a traffic accident result according to some embodiments of the present disclosure
- FIGs. 6A-6C are schematic diagrams illustrating exemplary traffic accidents according to some embodiments of the present disclosure.
- FIG. 7 is a schematic diagram illustrating an exemplary target image according to some embodiments of the present disclosure.
- FIG. 8 is a schematic diagram illustrating an exemplary process for adjusting a size of an image according to some embodiments of the present disclosure
- FIGs. 9A and 9B are schematic diagrams illustrating an exemplary process for determining a static region in a target image using a detection model according to some embodiments of the present disclosure
- FIGs. 10A-10C are schematic diagrams illustrating an exemplary process for determining a static region in a target image using a segmentation model according to some embodiments of the present disclosure
- FIG. 11 is a schematic diagram illustrating an exemplary 3D transformation result according to some embodiments of the present disclosure.
- FIG. 12 is a schematic diagram illustrating an exemplary first model according to some embodiments of the present disclosure.
- FIG. 13 is a block diagram illustrating an exemplary accident detection system according to some embodiments of the present disclosure.
- FIG. 14 is a flowchart illustrating an exemplary process for determining a traffic accident result according to some embodiments of the present disclosure
- FIG. 15 is a flowchart illustrating an exemplary process for extracting a static region in a target image according to some embodiments of the present disclosure
- FIG. 16 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure
- FIG. 17 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure
- FIG. 18 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure
- FIG. 19 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure
- FIG. 20 is a flowchart illustrating an exemplary process for modifying a confidence value associated with a traffic accident result according to some embodiments of the present disclosure
- FIG. 21 is a flowchart illustrating an exemplary process for modifying a confidence value associated with a traffic accident result according to some embodiments of the present disclosure
- FIG. 22 is a block diagram illustrating an exemplary electronic device according to some embodiments of the present disclosure.
- FIG. 23 is a block diagram illustrating an exemplary computer readable storage medium according to some embodiments of the present disclosure.
- the modules (or units, blocks, units) described in the present disclosure may be implemented as software and/or hardware modules and may be stored in any type of non-transitory computer-readable medium or other storage devices.
- a software module may be compiled and linked into an executable program. It will be appreciated that software modules can be callable from other modules or from themselves, and/or can be invoked in response to detected events or interrupts.
- Software modules configured for execution on computing devices can be provided on a computer readable medium, such as a compact disc, a digital video disc, a flash drive, a magnetic disc, or any other tangible medium, or as a digital download (and can be originally stored in a compressed or installable format that requires installation, decompression, or decryption prior to execution) .
- Such software code can be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device.
- Software instructions can be embedded in a firmware, such as an EPROM.
- hardware modules e.g., circuits
- circuits can be included of connected or coupled logic units, such as gates and flip-flops, and/or can be included of programmable units, such as programmable gate arrays or processors.
- the modules or computing device functionality described herein are preferably implemented as hardware modules, but can be software modules as well. In general, the modules described herein refer to logical modules that can be combined with other modules or divided into units despite their physical organization or storage.
- the flowcharts used in the present disclosure illustrate operations that systems implement according to some embodiments of the present disclosure. It is to be expressly understood, the operations of the flowcharts may be implemented not in order. Conversely, the operations may be implemented in inverted order, or simultaneously. Moreover, one or more other operations may be added to the flowcharts. One or more operations may be removed from the flowcharts.
- An aspect of the present disclosure relates to systems and methods for determining a traffic accident result based on a video of a target scene using a first model (also referred to as a traffic accident detection model) .
- the systems may determine a target image associated with the target scene based on the video.
- a dynamic region e.g., including one or more moving objects
- the systems may also extract a static region (e.g., including one or more static objects) associated with the target scene in the target image based on a second model (e.g., a detection model, a segmentation model) .
- the systems may determine the traffic accident result by processing the static region in the target image based on the first model.
- one or more objects associated with the traffic accident in the traffic scene may be in a static state while one or more objects (e.g., a vehicle, a driver) irrelevant with the traffic accident in the traffic scene may be in a motion state.
- the systems may first extract the static region and then determine the traffic accident result by merely taking the static region in the target image into consideration, thereby not only ensuring the accuracy of the traffic accident detection but also improving the efficiency of the traffic accident detection.
- FIG. 1 is a schematic diagram illustrating an exemplary accident detection system according to some embodiments of the present disclosure.
- the accident detection system 100 may include an image capture device 110, a processing device 120, one or more terminals 130, a network 140, and a storage device 150.
- the image capture device 110 may be configured to acquire an image (e.g., a traffic image) or a video (e.g., a traffic video) .
- the image or the video may be two-dimensional (2D) , three-dimensional (3D) , etc.
- the image capture device 110 may include a camera, a video camera, a mobile phone, a tablet, an e-book reader, or the like, or any combination thereof.
- a position and/or an orientation (e.g., an angle, a direction) of the image capture device 110 may be adjustable.
- the image capture device 110 may acquire image (s) continuously, within a specific time interval, or based on a control command.
- the acquired image (s) may be stored in the storage device 150, or transmitted to other components via the network 140 for storage or further processing.
- the image capture device 110 may communicate with one or more components of the accident detection system 100 (e.g., the processing device 120, the terminal (s) 130, the storage device 150) via the network 140. In some embodiments, the image capture device 110 may be directly connected to one or more components of the accident detection system 100 (e.g., the processing device 120, the terminal (s) 130, the storage device 150) .
- the processing device 120 may include a single server or a server group.
- the server group may be centralized or distributed (e.g., the processing device 120 may be a distributed system) .
- the processing device 120 may be local or remote.
- the processing device 120 may access information and/or data stored in the image capture device 110, the terminal (s) 130, and/or the storage device 150 via the network 140.
- the processing device 120 may be directly connected to the image capture device 110, the terminal (s) 130, and/or the storage device 150 to access stored information and/or data.
- the processing device 120 may be implemented on a cloud platform or an onboard computer.
- the cloud platform may include a private cloud, a public cloud, a hybrid cloud, a community cloud, a distributed cloud, an inter-cloud, a multi-cloud, or the like, or any combination thereof.
- the processing device 120 may be implemented on a computing device 200 including one or more components illustrated in FIG. 2 of the present disclosure.
- the processing device 120 may process information and/or data associated with traffic accident detection to perform one or more functions described in the present disclosure. For example, the processing device 120 may determine a target image associated with a target scene based on a video of the target scene. The target image may include a static region associated with the target scene and a dynamic region associated with the target scene. The processing device 120 may determine a result indicating whether a traffic accident happens in the static region by processing, based on a first model, the static region associated with the target scene in the target image. In some embodiments, the processing device 120 may include one or more processing engines (e.g., single-core processing engine (s) or multi-core processor (s) ) .
- processing engines e.g., single-core processing engine (s) or multi-core processor (s)
- the processing device 120 may include a central processing unit (CPU) , an application-specific integrated circuit (ASIC) , an application-specific instruction-set processor (ASIP) , a graphics processing unit (GPU) , a physics processing unit (PPU) , a digital signal processor (DSP) , a field-programmable gate array (FPGA) , a programmable logic device (PLD) , a controller, a microcontroller unit, a reduced instruction-set computer (RISC) , a microprocessor, or the like, or any combination thereof.
- the image capture device 110 may be part of the processing device 120.
- the processing device 120 may be connected to the network 140 to communicate with one or more components (e.g., the image capture device 110, the terminal (s) 130, the storage device 150) of the accident detection system 100. In some embodiments, the processing device 120 may be directly connected to or communicate with one or more components (e.g., the image capture device 110, the terminal (s) 130, the storage device 150) of the accident detection system 100.
- the terminal (s) 130 may be configured to receive/transmit data and/or information from the image capture device 110, the processing device 120, and/or the storage device 150.
- the user terminal (s) 130 may receive a traffic accident result from the processing device 120.
- the user terminal (s) 130 may transmit an instruction to the image capture device 110 and/or the processing device 120.
- the terminal (s) 130 may include a tablet computer 130-1, a laptop computer 130-2, a built-in device in a vehicle 130-3, a mobile device 130-4, or the like, or any combination thereof.
- the terminal (s) 130 may communicate with one or more components of the accident detection system 100 (e.g., the image capture device 110, the processing device 120, the storage device 150) via the network 140. In some embodiments, the terminal (s) 130 may be directly connected to one or more components of the accident detection system 100 (e.g., the image capture device 110, the processing device 120, the storage device 150) .
- the network 140 may facilitate exchange of information and/or data.
- one or more components e.g., the image capture device 110, the processing device 120, the terminal (s) 130, the storage device 150
- the processing device 120 may obtain a video of a target scene from the storage device 150 via the network 140.
- the network 140 may be any type of wired or wireless network, or a combination thereof.
- the network 140 may include a cable network, a wireline network, an optical fiber network, a telecommunications network, an intranet, an Internet, a local area network (LAN) , a wide area network (WAN) , a wireless local area network (WLAN) , a metropolitan area network (MAN) , a public telephone switched network (PSTN) , a Bluetooth network, a ZigBee network, a near field communication (N FC) network, or the like, or any combination thereof.
- the network 140 may include one or more network access points.
- the network 140 may include wired or wireless network access points (e.g., a point 140-1, a point 140-2) , through which one or more components of the accident detection system 100 may be connected to the network 140 to exchange data and/or information.
- the storage device 150 may store data and/or instructions.
- the storage device 150 may store data obtained from the image capture device 110, the processing device 120, the terminal (s) 130, or an external storage device.
- the storage device 150 may store a video of a target scene acquired by the image capture device 110.
- the storage device 150 may store data and/or instructions that the processing device 120 may execute or use to perform exemplary methods described in the present disclosure.
- the storage device 150 may store instructions that the processing device 120 may execute or use to determine a result indicating whether a traffic accident happens in a static region by processing, based on a first model, the static region associated with a target scene in a target image.
- the storage device 150 may include a mass storage, a removable storage, a volatile read-and-write memory, a read-only memory (ROM) , or the like, or any combination thereof.
- Exemplary mass storage may include a magnetic disk, an optical disk, a solid-state drive, etc.
- Exemplary removable storage may include a flash drive, a floppy disk, an optical disk, a memory card, a zip disk, a magnetic tape, etc.
- Exemplary volatile read-and-write memory may include a random access memory (RAM) .
- Exemplary RAM may include a dynamic RAM (DRAM) , a double date rate synchronous dynamic RAM (DDR SDRAM) , a static RAM (SRAM) , a thyristor RAM (T-RAM) , and a zero-capacitor RAM (Z-RAM) , etc.
- Exemplary ROM may include a mask ROM (MROM) , a programmable ROM (PROM) , an erasable programmable ROM (EPROM) , an electrically-erasable programmable ROM (EEPROM) , a compact disk ROM (CD-ROM) , and a digital versatile disk ROM, etc.
- the storage device 150 may be implemented on a cloud platform.
- the cloud platform may include a private cloud, a public cloud, a hybrid cloud, a community cloud, a distributed cloud, an inter-cloud, a multi-cloud, or the like, or any combination thereof.
- the storage device 150 may be connected to the network 140 to communicate with one or more components (e.g., the image capture device 110, the processing device 120, the terminal (s) 130) of the accident detection system 100.
- One or more components of the accident detection system 100 may access the data or instructions stored in the storage device 150 via the network 140.
- the storage device 150 may be directly connected to or communicate with one or more components (e.g., the image capture device 110, the processing device 120, the terminal (s) 130) of the accident detection system 100.
- the storage device 150 may be part of the processing device 120.
- the storage device 150 may be integrated into the processing device 120.
- two or more components of the accident detection system 100 may be integrated in a same device.
- the image capture device 110, the processing device 120, and the storage device 150 may be integrated into a same device (e.g., a camera, a smartphone, a computer, a workstation, a server) .
- one or more components of the accident detection system 100 may be remote from each other.
- the image capture device 110 may be remote from the processing device 120.
- accident detection system 100 is merely provided for the purposes of illustration, and is not intended to limit the scope of the present disclosure.
- multiple variations or modifications may be made under the teachings of the present disclosure. However, those variations and modifications do not depart from the scope of the present disclosure.
- the accident detection system 100 may also include a user device (not shown) configured to receive information and/or data from the image capture device 110, the processing device 120, the terminal (s) 130, and/or the storage device 150.
- the user device may provide a user interface via which a user may view information (e.g., a traffic accident result) and/or input data (e.g., a video of a target scene) and/or instructions to the accident detection system 100.
- the accident detection system 100 may also include an input/output (I/O) component.
- the I/O component may be configured to input or output signals, data, or information.
- the I/O component may include an input device and an output device.
- Exemplary input devices may include a keyboard, a mouse, a touch screen, a microphone, or the like, or any combination thereof.
- Exemplary output devices may include a display device, a speaker, a printer, a projector, or the like, or any combination thereof.
- FIG. 2 is a schematic diagram illustrating exemplary hardware and/or software components of an exemplary computing device according to some embodiments of the present disclosure.
- the computing device 200 may be used to implement any component of the accident detection system 100 as described herein.
- the processing device 120 may be implemented on the computing device 200, via its hardware, software program, firmware, or a combination thereof.
- only one such computer is shown, for convenience, the computer functions relating to traffic accident detection as described herein may be implemented in a distributed fashion on a number of similar platforms to distribute the processing load.
- the computing device 200 may include COM ports 250 con nected to and from a network connected thereto to facilitate data communications.
- the computing device 200 may also include a processor (e.g., a processor 220) , in the form of one or more processors (e.g., logic circuits) , for executing program instructions.
- the processor 220 may include interface circuits and processing circuits therein.
- the interface circuits may be configured to receive electronic signals from a bus 210, wherein the electronic signals encode structured data and/or instructions for the processing circuits to process.
- the processing circuits may conduct logic calculations, and then determine a conclusion, a result, and/or an instruction encoded as electronic signals. Then the interface circuits may send out the electronic signals from the processing circuits via the bus 210.
- the computing device 200 may further include one or more storages configured to store various data files (e.g., program instructions) to be processed and/or transmitted by the computing device 200.
- the one or more storages may include a high speed random access memory (not shown) , a non-volatile memory (e.g., a magnetic storage device, a flash memory, or other non-volatile solid state memories) (not shown) , a disk 270, a read-only memory (ROM) 230, a random-access memory (RAM) 240, or the like, or any combination thereof.
- the one or more storages may further include a remote storage corresponding to the processor 220. The remote storage may connect to the computing device 200 via the network 140.
- the computing device 200 may also include program instructions stored in the one or more storages (e.g., the ROM 230, RAM 240, and/or another type of non-transitory storage medium) to be executed by the processor 220.
- the methods and/or processes of the present disclosure may be implemented as the program instructions.
- the computing device 200 may also include an I/O component 260, supporting input/output between the computing device 200 and other components.
- the computing device 200 may also receive programming and data via network communications.
- processors 220 are also contemplated; thus, operations and/or method steps performed by one processor 220 as described in the present disclosure may also be jointly or separately performed by the multiple processors.
- the processor 220 of the computing device 200 executes both operation A and operation B, it should be understood that operation A and operation B may also be performed by two different processors 220 jointly or separately in the computing device 200 (e.g., a first processor executes operation A and a second processor executes operation B, or the first and second processors jointly execute operations A and B) .
- FIG. 3 is a schematic diagram illustrating exemplary hardware and/or software components of an exemplary mobile device according to some embodiments of the present disclosure.
- the processing device 120 or the user device may be implemented on the mobile device 300.
- the mobile device 300 may include a communication platform 310, a display 320, a graphics processing unit (GPU) 330, a central processing unit (CPU) 340, an I/O 350, a memory 360, a mobile operating system (OS) 370, and a storage 390.
- any other suitable components including but not limited to a system bus or a controller (not shown) , may also be in the mobile device 300.
- the mobile operating system 370 e.g., iOS TM , Android TM , Windows Phone TM
- the applications 380 may include a browser or any other suitable mobile apps for receiving and rendering information relating to traffic accident detection or other information from the accident detection system 100.
- User interactions with the information stream may be achieved via the I/O 350 and provided to the processing device 120 and/or other components of the accident detection system 100 via the network 140.
- FIG. 4 is a block diagram illustrating an exemplary processing device according to some embodiments of the present disclosure.
- the processing device 120 may include a video obtainment module 410, an image determination module 420, a region determination module 430, and a detection module 440.
- the video obtainment module 410 may be configured to obtain a video of a target scene (e.g., a traffic scene) that is captured by an image capture device (e.g., the image capture device 110 in FIG. 1) .
- the video may be acquired by the image capture device in real-time.
- the video may include a plurality of images of the target scene.
- the image determination module 420 may be configured to determine a target image (e.g., image 710 in FIG. 7) associated with the target scene based on the video. In some embodiments, the image determination module 420 may obtain a first image and one or more second images from the video. The one or more second images may be captured prior to the first image. The image determination module 420 may determine the target image by processing a dynamic region associated with the target scene in the first image based on the one or more second images. In some embodiments, the dynamic region may include one or more moving objects, for example, a moving vehicle, a moving pedestrian, a moving driver, etc., in the target scene.
- the image determination module 420 may generate a fusion result by fusing the first image and the one or more second images and determine the target image based on the fusion result. In some embodiments, the image determination module 420 may blur the dynamic region in the target image based on the fusion result, that is, the target image may include the blurred dynamic region (e.g., a region 711 in FIG. 7) in the target scene.
- the image determination module 420 may blur the dynamic region in the target image based on the fusion result, that is, the target image may include the blurred dynamic region (e.g., a region 711 in FIG. 7) in the target scene.
- the image determination module 420 may adjust a size (e.g., an area, a length, a width, an aspect ratio) of the first image and/or the one or more second images, and then determine the target image based on the adjusted first image and/or the adjusted one or more second images accordingly.
- the image determination module 420 may adjust the size of the first image and/or the one or more second images using a size adjustment technique.
- the size adjustment technique may include a secondary nonlinear interpolation technique, a linear interpolation technique, etc.
- the image determination module 420 may adjust the size of the first image and/or the one or more second images to a reference size.
- the reference size may be predetermined according to practical demands, for example, an accuracy of traffic accident detection, a type of a first model used to determine the traffic accident, a type of a second model used to determine a static region, etc.
- the image determination module 420 may enlarge a length of a short side of an image (e.g., the first image, the one or more second images) .
- an enlarged portion of the image may be filled with a color whose pixel value is a specific value, e.g., 127.
- the region determination module 430 may be configured to may determine a static region associated with the target scene in the target image. In some embodiments, the region determination module 430 may determine one or more regions associated with one or more static objects (e.g., a vehicle, a pedestrian) of the target scene in the target image by processing the target image (e.g., a portion of the target image excluding the dynamic region) based on a second model (e.g., a neural network model) . As used herein, a static object may be stationary relative to a road surface.
- a static object may be stationary relative to a road surface.
- the region determination module 430 may also determine a position associated with one of the one or more regions (e.g., a coordinate of a center of a region) , a size (e.g., a length, a width, an area) of one of the one or more regions, etc., based on the second model.
- a position associated with one of the one or more regions e.g., a coordinate of a center of a region
- a size e.g., a length, a width, an area
- the second model may include a detection model (e.g., you-only-look-once (YOLO) neural network model) .
- the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more bounding boxes (e.g., rectangles 2, 3, or 4 in FIG. 9A) (e.g., a two-dimensional bounding box) of the one or more static objects determined by the detection model.
- each of the one or more bounding boxes may correspond to a static object.
- the second model may include a segmentation model.
- the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more connected components (e.g., regions 1030 and 1040 in FIG. 10B) of the one or more static objects determined by the segmentation model.
- each of the one or more connected components may correspond to a type (e.g., a vehicle type (e.g., region 1030 in FIG. 10B) , a person type (e.g., region 1040 in FIG. 10B)) of the one or more static objects.
- the region determination module 430 may determine the static region associated with the target scene in the target image based on the one or more regions (e.g., the one or more bounding boxes, the one or more connected components) associated with the one or more static objects of the target scene in the target image. In some embodiments, the region determination module 430 may designate each of the one or more regions as one static region associated with the target scene in the target image. As used herein, a static region may refer to a minimum region containing one of the one or more static objects.
- the region determination module 430 may determine at least one of the one or more regions that satisfies a first condition.
- the first condition may relate to at least one distance between the one or more regions (e.g., one or more centers thereof) .
- the first condition may include that at least one distance between the at least one region is smaller than or equal to a distance threshold.
- the first condition may include that a ratio of a distance between two regions of the at least one region and a width of an object (e.g., a vehicle) inside one of the two regions is smaller than or equal to a ratio threshold (e.g., 1.2) .
- a ratio threshold e.g., 1.2
- the region determination module 430 may generate a combination result by combining the at least one region.
- the combination result may include a combined bounding box (e.g., rectangle 6 in FIG. 9B, rectangle f in FIG. 10C) contained the at least one region.
- the region determination module 430 may determine the static region associated with the target scene in the target image based on the combination result.
- the region determination module 430 may designate a region (e.g., a minimum region containing the combined bounding box) corresponding to the combined bounding box as one static region.
- the region determination module 430 may designate each region of a remaining portion of the one or more regions as one static region.
- the remaining portion of the one or more regions may be probably irrelevant with a traffic accident.
- the remaining portion of the one or more regions may include road plants 712 in FIG. 7 which are irrelevant with a traffic accident.
- the region determination module 430 may exclude at least a portion of the remaining portion of the one or more regions from the static region.
- the region determination module 430 may determine whether a distance between a region of the remaining portion of the one or more regions and one of the at least one region is larger than a second distance threshold.
- the region determination module 430 may exclude the region from the static region, that is the region determination module 430 may not designate the region as one static region; in response to determining that the distance is smaller than or equal to the second distance threshold, the region determination module 430 may designate the region as one static region.
- the second distance threshold may be larger than the distance threshold.
- the region determination module 430 may enlarge the bounding box (or the combined bounding box) by a preset ratio and designate a region corresponding to the enlarged bounding box (e.g., rectangle 6’ in FIG. 9A) as the static region.
- the ratio may be preset according to practical demands, for example, a type of the target scene, a size (e.g., a length, a width, an area, an aspect ratio) of the bounding box (or the combined bounding box) , a size (e.g., a length, a width, an area, an aspect ratio) of an object inside the bounding box (or the combined bounding box) .
- the preset ratio may be 1.2.
- the preset ratio may be 0.5 of the width of a vehicle inside the bounding box.
- the detection module 440 may be configured to determine a result (also referred to as a traffic accident result) indicating whether a traffic accident happens in the static region by processing, based on a first model, the static region associated with the target scene in the target image.
- the traffic accident result may also include a confidence value associated with the traffic accident, a type of the traffic accident, etc.
- the detection module 440 may modify the confidence value associated with the traffic accident. In some embodiments, in response to determining that the confidence value fails to satisfy a second condition (e.g., larger than the value threshold (e.g., 0.5) ) , the detection module 440 may determine whether there are at least two of the one or more regions associated with the one or more static objects in the static region that overlap.
- a second condition e.g., larger than the value threshold (e.g., 0.5)
- the detection module 440 may modify the confidence associated with the traffic accident; in response to determining that there are no at least two of the one or more regions associated with the one or more static objects in the static region that overlap, the detection module 440 may not modify the confidence value associated with the traffic accident and output a result indicating that there is no traffic accident happens in the static region. In response to determining that the confidence value satisfies the second condition, the detection module 440 may output a result indicating that a traffic accident happens in the static region.
- the one or more regions associated with the one or more static objects may be two-dimensional.
- the detection module 440 may determine one or more three-dimensional (3D) regions (e.g., rectangles 1110 and 1120 in FIG. 11) associated with the one or more static objects in the static region by performing a 3D transformation on the one or more 2D regions associated with the one or more static objects based on one or more coordinates associated with the one or more 2D regions (e.g., a coordinate of a vertex or a central point of a 2D region) .
- the detection module 440 may modify the confidence value associated with the traffic accident based on an overlap (also referred to as an overlapping region) between the one or more 3D regions associated with the one or more static objects in the static region.
- the detection module 440 may determine a ratio between a volume of the overlap between the one or more 3D regions and a volume of one of the one or more 3D regions (e.g., a minimum volume of the one or more 3D regions) and modify the confidence value based on the ratio.
- the detection module 440 may determine the modified confidence value by performing a weighted summation on the ratio and the confidence value. For example, the detection module 440 may designate a value corresponding to the weighted summation result as the modified confidence value.
- the detection module 440 may determine that the traffic accident happens in the static region. In some embodiments, in response to determining that the modified confidence value is smaller than or equal to the value threshold, the detection module 440 may determine there is no traffic accident in the static region.
- the modules in the processing device 120 may be connected to or communicated with each other via a wired connection or a wireless connection.
- the wired connection may include a metal cable, an optical cable, a hybrid cable, or the like, or any combination thereof.
- the wireless connection may include a Local Area Network (LAN) , a Wide Area Network (WAN) , a Bluetooth, a ZigBee, a Near Field Communication (NFC) , or the like, or any combination thereof.
- LAN Local Area Network
- WAN Wide Area Network
- Bluetooth a ZigBee
- NFC Near Field Communication
- the image determination module 420 and the region determination module 430 may be combined as a single module which may determine the target image associated with the target scene based on the video and determine the static region associated with the target scene in the target image.
- the processing device 120 may include a storage module (not shown) which may be used to store data generated by the above-mentioned modules.
- FIG. 5 is a flowchart illustrating an exemplary process for determining a traffic accident result according to some embodiments of the present disclosure.
- FIGs. 6A-6C are schematic diagrams illustrating exemplary traffic accidents according to some embodiments of the present disclosure.
- FIG. 7 is a schematic diagram illustrating an exemplary target image according to some embodiments of the present disclosure.
- FIG. 8 is a schematic diagram illustrating an exemplary process for adjusting a size of an image according to some embodiments of the present disclosure.
- FIGs. 9A and 9B are schematic diagrams illustrating an exemplary process for determining a static region in a target image using a detection model according to some embodiments of the present disclosure.
- FIGS. 10A-10C are schematic diagrams illustrating an exemplary process for determining a static region in a target image using a segmentation model according to some embodiments of the present disclosure.
- FIG. 11 is a schematic diagram illustrating an exemplary 3D transformation result according to some embodiments of the present disclosure.
- the process 500 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 500.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 500 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 5 and described below is not intended to be limiting.
- the processing device 120 e.g., the video obtainment module 410) (e.g., the interface circuits of the processor 220) may obtain a video of a target scene (e.g., a traffic scene) that is captured by an image capture device (e.g., the image capture device 110 in FIG. 1) .
- the video may be acquired by the image capture device in real-time.
- the video may include a plurality of images of the target scene, for example, arranged in a time order.
- a time interval between two adjacent images of the plurality of images may be set according to practical demands.
- the time interval between two adjacent images may relate to a type of the target scene (e.g., a traffic flow of the target scene, a speed limit of the target scene) , an accuracy of traffic accident detection, a speed of traffic accident detection, etc.
- a time interval (e.g., 5 frames) between two adjacent images corresponding to the first road may be smaller than or equal to a time interval (10 frames) between two adjacent images corresponding to the second road.
- a time interval (e.g., 5 frames) between two adjacent images corresponding to a tunnel may be smaller than or equal to a time interval (10 frames) between two adjacent images corresponding to a city road.
- the processing device 120 may determine a target image (e.g., image 710 in FIG. 7) associated with the target scene based on the video.
- the processing device 120 may obtain a first image and one or more second images from the video. The one or more second images may be captured prior to the first image.
- the processing device 120 may determine the target image by processing a dynamic region associated with the target scene in the first image based on the one or more second images.
- the dynamic region may include one or more moving objects, for example, a moving vehicle, a moving pedestrian, a moving driver, etc., in the target scene.
- the dynamic region may include a minimum region containing the one or more moving objects.
- the processing device 120 may determine the target image using an exponential smoothing and filtering technique, a Gaussian filtering technique, etc.
- the processing device 120 may perform traffic accident detection for each of the plurality images in the video, and the first image may be each of the plurality images captured in a current time point by the image capture device.
- the processing device 120 may generate a fusion result by fusing the first image and the one or more second images and determine the target image based on the fusion result.
- the processing device 120 may blur the dynamic region in the target image based on the fusion result, that is, the target image may include the blurred dynamic region (e.g., a region 711 in FIG. 7) in the target scene.
- the processing device 120 may determine the target image according to formula (1) below:
- B refers to the target image
- B′ refers to the one or more second images before the blurred operation and/or the fusion operation
- P refers to the first image before the blurred operation and/or the fusion operation
- ⁇ refers to an exponential smoothing weight associated with the first image, which is used to control a reserving degree of information of the one or more second images after the blurred operation and/or the fusion operation.
- the processing device 120 may set the exponential smoothing weight associated with the first image according to practical demands.
- the exponential smoothing weight may be 0.15.
- the one or more second images may include all images captured before the first image and included in the video. In some embodiments, the one or more second images may include a preset count (e.g., 5, 10, 15) of images captured before the first image and included in the video. A time interval between two adjacent second images may be multiple of the time interval between two adjacent images of the plurality of images included in the video. In some embodiments, the processing device 120 may determine the target image by blurring the dynamic region in the first image directly using all of the one or more second images.
- the processing device 120 may determine the target image by blurring a dynamic region in an intermediate image using at least one of the one or more second images captured before the intermediate image (e.g., one of the one or more second images captured before the first image and next to the first image) and then blurring the dynamic region in the first image using the blurred intermediate image.
- the one or more second images and the first image may be arranged in a time order.
- a time interval between two adjacent images among the one or more second images and the first image may be the same as the time interval between two adjacent images among the plurality of images.
- the time interval between two adjacent images among the one or more second images and the first image may be multiple of the time interval between two adjacent images among the plurality of images.
- the time interval between two adjacent images among the one or more second images and the first image may relate to a type of the target scene (e.g., a traffic flow of the target scene, a speed limit of the target scene) , an accuracy of traffic accident detection, a speed of traffic accident detection, etc.
- a type of the target scene e.g., a traffic flow of the target scene, a speed limit of the target scene
- an accuracy of traffic accident detection e.g., a speed of traffic accident detection, etc.
- the processing device 120 may adjust a size (e.g., an area, a length, a width, an aspect ratio) of the first image and/or the one or more second images, and then determine the target image based on the adjusted first image and/or the adjusted one or more second images accordingly.
- the processing device 120 may adjust the size of the first image and/or the one or more second images using a size adjustment technique.
- the size adjustment technique may include a secondary linear interpolation technique, a nonlinear interpolation technique, etc.
- the processing device 120 may adjust the size of the first image and/or the one or more second images to a reference size.
- the reference size may be predetermined according to practical demands, for example, an accuracy of traffic accident detection, a type of a first model used to determine the traffic accident, a type of a second model used to determine a static region, etc.
- the processing device 120 may enlarge a length of a short side of an image (e.g., the first image, the one or more second images) .
- an enlarged portion of the image may be filled with a color whose pixel value is a specific value within a range of 0 and 255, e.g., 0, 127, 255.
- the specific value may be default of the accident detection system 100 or predetermined according to practical demands.
- the processing device 120 may enlarge the length of the short side of the image to be the same as the length of the long side of the image. Take an image 810 in FIG.
- a length of a long side of the image 810 is a
- a length of a short side of the image 810 is b smaller than a.
- the processing device 120 may enlarge the length of the short side of the image 810 to c (same as or different from a) by filling an enlarged portion of the image 810 with a color whose pixel value is 127. Since the target image that is directly determined based on the first image and/or the one or more second images after the size adjustment may deform, affecting the accuracy of the traffic accident detection, the processing device 120 may first enlarge the length of the short side of the image and then perform the size adjustment of the image.
- the processing device 120 may determine a static region associated with the target scene in the target image.
- the processing device 120 may determine one or more regions associated with one or more static objects (e.g., a vehicle, a pedestrian) of the target scene in the target image by processing the target image (e.g., a portion of the target image excluding the dynamic region) based on a second model (e.g., a neural network model) .
- a static object may be stationary relative to a road surface.
- the processing device 120 may determine the static region without identifying objects inside the dynamic region, thereby reducing a computing load for determining the static region and improving an accuracy for determining the static region.
- the processing device 120 may also determine a position associated with one of the one or more regions (e.g., a coordinate of a center of a region) , a size (e.g., a length, a width, an area) of one of the one or more regions, etc., based on the second model.
- the second model may include a detection model (e.g., you-only-look-once (YOLO) neural network model) .
- the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more bounding boxes (e.g., rectangles 2, 3, or 4 in FIG. 9A) (e.g., a two-dimensional bounding box) of the one or more static objects determined by the detection model.
- each of the one or more bounding boxes may correspond to a static object.
- the second model may include a segmentation model.
- the one or more regions associated with the one or more static objects of the target scene in the target image may include one or more connected components (e.g., regions 1030 and 1040 in FIG. 10B) of the one or more static objects determined by the segmentation model.
- each of the one or more connected components may correspond to a type (e.g., a vehicle type (e.g., region 1030 in FIG. 10B) , a person type (e.g., region 1040 in FIG. 10B)) of the one or more static objects.
- a connected component of a static object determined based on the segmentation model may be a contour of the static object.
- a bounding box of a static object determined based on the detection model may be a maxium bounding rectangle of the static object. Therefore, the extraction of the static region based on the detection model may be less time-comsuming than the extraction of the static region based on the segmentation model. An accuracy for extracting the static region based on the segmentation model may be higher than an accuracy for extracting the static region based on the detection model.
- the second model (e.g., the detection model, the segmentation model) may be trained using a plurality of sample images of various sample traffic accidents, for example, a front collision between sample vehicles, a front-end collision between sample vehicles (e.g., as shown in FIG. 6A) , a side collision between sample vehicles, a collision between a sample vehicle and a sample fixture (e.g., as shown in FIG. 6B) , a side flip of a sample vehicle (e.g., 620 shown in FIG. 6C) , a disintegration of a sample vehicle, or the like, or any combination thereof.
- a front collision between sample vehicles e.g., a front-end collision between sample vehicles (e.g., as shown in FIG. 6A)
- a side collision between sample vehicles e.g., a collision between a sample vehicle and a sample fixture
- a side flip of a sample vehicle e.g., 620 shown in FIG. 6C
- each of the plurality of sample images may have a label as a known output (e.g., a ground truth) .
- a label of a sample image may indicate a bounding box or a connected component of a static object in the sample image, and/or a type of the static object in the sample image.
- the values of the second model’s parameters may be determined or updated by performing an iteration of a backpropagation training procedure, e.g., a stochastic gradient descent backpropagation training technique.
- the training process may backpropagate the error (e.g., indicated by a loss function) determined for the output of the second model by comparing the output of the second model with the label (e.g., the ground truth) .
- the training process may terminate until the error satisfies a condition (e.g., the error smaller than or equal to the threshold) .
- the processing device 120 may determine the static region associated with the target scene in the target image based on the one or more regions (e.g., the one or more bounding boxes, the one or more connected components) associated with the one or more static objects of the target scene in the target image. In some embodiments, the processing device 120 may designate each of the one or more regions as one static region associated with the target scene in the target image.
- a static region may refer to a minimum region containing one of the one or more static objects.
- the processing device 120 may determine at least one of the one or more regions that satisfies a first condition.
- the first condition may relate to at least one distance between the one or more regions (e.g., one or more centers thereof) .
- the first condition may include that at least one distance between the at least one region is smaller than or equal to a distance threshold.
- the first condition may include that a ratio of a distance between two regions of the at least one region and a width of an object (e.g., a vehicle) inside one of the two regions is smaller than or equal to a ratio threshold (e.g., 1.2) .
- a ratio threshold e.g., 1.2
- the processing device 120 may generate a combination result by combining the at least one region.
- the combination result may include a combined bounding box (e.g., rectangle 6 in FIG. 9B, rectangle f in FIG. 10C) contained the at least one region.
- the processing device 120 may determine the static region associated with the target scene in the target image based on the combination result.
- the processing device 120 may designate a region (e.g., a minimum region containing the combined bounding box) corresponding to the combined bounding box as one static region.
- the processing device 120 may designate each region of a remaining portion of the one or more regions as one static region.
- the processing device 120 may determine the combined bounding box based on a disjoint set union (DSU) structure.
- DSU disjoint set union
- the processing device 120 may exclude at least a portion of the remaining portion of the one or more regions from the static region. In some embodiments, the processing device 120 may determine whether a distance between a region of the remaining portion of the one or more regions and one of the at least one region is larger than a second distance threshold.
- the processing device 120 may exclude the region from the static region, that is the processing device 120 may not designate the region as one static region; in response to determining that the distance is smaller than or equal to the second distance threshold, the processing device 120 may designate the region as one static region.
- the second distance threshold may be larger than the distance threshold. In such cases, the blurred dynamic region and the at least a portion of the remaining portion may be unnecessary for further traffic accident detection, thereby not only reserving information that needs for the following traffic accident detection process but also reducing a computing load for the following traffic accident detection process.
- the target image may include five static objects, i.e., a first object, a second object, a third object, a fourth object, and a fifth object, bounding boxes of which are a first bounding box, a second bounding box, a third bounding box, a fourth bounding box, and a fifth bounding box determined by the detection model.
- the processing device 120 may generate a sixth bounding box by combining the second bounding, the third bounding box, and the fourth bounding box.
- the target image may include a static region corresponding to the first bounding box, a static region corresponding to the fifth bounding box, and a static region corresponding to the sixth bounding box.
- the target image may include two static vehicles 1010 and a static person 1020.
- Region 1030 indicating a first connected component corresponding to the two static vehicles 1010 and region 1040 indicating a second connected component corresponding to the static person 1020 may be determined based on the segmentation model. Since the first connected component and the second connected component satisfy the first condition, a combined bounding box f may be determined by combining the first connected component and the second connected component.
- the processing device 120 may designate a region corresponding to the combined bounding box f as one static region.
- the processing device 120 may enlarge the bounding box (or the combined bounding box) by a preset ratio and designate a region corresponding to the enlarged bounding box (e.g., rectangle 6’ in FIG. 9A) as the static region.
- the ratio may be preset according to practical demands, for example, a type of the target scene, a size (e.g., a length, a width, an area, an aspect ratio) of the bounding box (or the combined bounding box) , a size (e.g., a length, a width, an area, an aspect ratio) of an object inside the bounding box (or the combined bounding box) .
- the preset ratio may be 1.2.
- the preset ratio may be 0.5 of the width of a vehicle inside the bounding box.
- the processing device 120 may determine a result (also referred to as a traffic accident result) indicating whether a traffic accident happens in the static region by processing, based on a first model, the static region associated with the target scene in the target image.
- the traffic accident result may also include a confidence value associated with the traffic accident, a type of the traffic accident, etc.
- the larger the confidence value is, the larger probability that the traffic accident exits in the static region may be.
- the confidence value if the confidence value exceeds a value threshold (e.g., 0.5) , the confidence value may be encoded as “1” ; if the confidence value is smaller than or equal to the value threshold, the confidence value may be encoded as “0” .
- a value threshold e.g., 0.5
- the confidence value may be encoded as “1” ; if the confidence value is smaller than or equal to the value threshold, the confidence value may be encoded as “0” .
- “1” may indicate that there is a traffic accident in the target image; “0” may indicate that there is no traffic accident in the target image.
- the type of the traffic accident may include a collision between vehicles, a collision between a vehicle and a preset fixture, a traffic accident of by a single vehicle, or the like, or any combination thereof.
- the type of the traffic accident may include a front collision between vehicles, a front-end collision between vehicles (e.g., as shown in FIG. 6A) , a side collision between vehicles, a collision between a vehicle and a preset fixture (e.g., as shown in FIG. 6B) , a side flip of a vehicle (e.g., 620 shown in FIG. 6C) , a disintegration of a vehicle, or the like, or any combination thereof.
- the preset fixture may include a road facility such as a road barrier (e.g., 610 shown in FIG. 6B) , a lamp rod, a marker rod, a building, a green plant, etc.
- a road barrier e.g., 610 shown in FIG. 6B
- the side flip or the disintegration of the vehicle may be caused by a flat tire of the vehicle, aging of a mechanical component of the vehicle, malfunction of a mechanical component of the vehicle, etc.
- the processing device 120 may adjust a size of the target image according to the first model, and determine the traffic accident result based on the adjusted target image accordingly. For example, the processing device 120 may adjust the size of the target image to be the same as a size of an image that the first model can process.
- the process for adjusting a size of the target image according to the first model may be the same as the process for adjusting the size of the first image and/or the one or more second images, the descriptions of which are not repeated.
- the first model may be trained using a plurality of sample images of various sample traffic accidents, for example, a front collision between sample vehicles, a front-end collision between sample vehicles (e.g., as shown in FIG. 6A) , a side collision between sample vehicles, a collision between a sample vehicle and a sample fixture (e.g., as shown in FIG. 6B) , a side flip of a sample vehicle (e.g., 620 shown in FIG. 6C) , a disintegration of a sample vehicle, or the like, or any combination thereof. More descriptions of a structure of the first model may be found elsewhere in the present disclosure, for example, FIG. 12 and the descriptions thereof.
- the first model may be trained using the plurality of sample images each having a label as a known output (e.g., a ground truth) .
- a label of a sample image may indicate a confidence value associated with the traffic accident of a static region in the sample image, and/or a type of the traffic accident of the sample image.
- the values of the first model’s parameters may be determined or updated by performing an iteration of a backpropagation training procedure, e.g., a stochastic gradient descent backpropagation training technique.
- the training process may backpropagate the error (e.g., indicated by a loss function) determined for the output of the first model by comparing the output of the first model with the label (e.g., the ground truth) .
- the training process may terminate until the error satisfies a condition (e.g., the error smaller than or equal to the threshold) .
- the processing device 120 may modify the confidence value associated with the traffic accident. In some embodiments, in response to determining that the confidence value fails to satisfy a second condition (e.g., larger than the value threshold (e.g., 0.5) ) , the processing device 120 may determine whether there are at least two of the one or more regions associated with the one or more static objects in the static region that overlap.
- a second condition e.g., larger than the value threshold (e.g., 0.5)
- the processing device 120 may modify the confidence associated with the traffic accident; in response to determining that there are no regions associated with the one or more static objects in the static region that overlap, the processing device 120 may not modify the confidence value associated with the traffic accident and output a result indicating that there is no traffic accident happens in the static region. In response to determining that the confidence value satisfies the second condition, the processing device 120 may output a result indicating that a traffic accident happens in the static region.
- the one or more regions associated with the one or more static objects may be two-dimensional.
- the processing device 120 may determine one or more three-dimensional (3D) regions (e.g., rectangles 1110 and 1120 in FIG. 11) associated with the one or more static objects in the static region by performing a 3D transformation on the one or more 2D regions associated with the one or more static objects based on one or more coordinates associated with the one or more 2D regions (e.g., a coordinate of a vertex or a central point of a 2D region) .
- the processing device 120 may determine the 3D regions according to formula (2) below:
- K refers to an intrinsic parameter (e.g., a focal length, a distortion parameter, etc. ) of the image capture device
- (x, y) refers to a coordinate of a point (e.g., a vertex, a central point) of the one or more regions in a 2D coordinate system (e.g., an image coordinate system)
- (X, Y, Z) refers to a coordinate of the point of the one or more regions in a 3D coordinate system (e.g., a world coordinate system)
- R refers to a rotation transformation vector between the 2D coordinate system and the 3D coordinate system
- t refers to a translation transformation vector between the 2D coordinate system and the 3D coordinate system.
- the processing device 120 may determine the projection transform matrix based on a process for determining a mapping relation between the 2D coordinate system and the 3D coordinate system, e.g., a least-squares technique, a Hough transformation technique, or a combination thereof.
- the processing device 120 may modify the confidence value associated with the traffic accident based on an overlap (also referred to as an overlapping region) between the one or more 3D regions associated with the one or more static objects in the static region.
- the processing device 120 may determine a ratio between a volume of the overlap between the one or more 3D regions and a volume of one of the one or more 3D regions (e.g., a minimum volume of the one or more 3D regions) and modify the confidence value based on the ratio.
- the processing device 120 may determine the modified confidence value by performing a weighted summation on the ratio and the confidence value. For example, the processing device 120 may designate a value corresponding to the weighted summation result as the modified confidence value.
- the processing device 120 may determine that the traffic accident happens in the static region. In some embodiments, in response to determining that the modified confidence value is smaller than or equal to the value threshold, the processing device 120 may determine there is no traffic accident in the static region.
- the confidence value may be determined by considering 2D information of the target scene, causing inaccurate traffic accident detection (e.g., failing to detect a slight collision between vehicles) .
- Information of the target scene may be enriched (e.g., 3D information compared to the 2D information) by determining the one or more 3D regions associated with the one or more static objects in the static region, such that the modified confidence value may be more accurate than the confidence value, thereby improving an accuracy for the traffic accident detection.
- a count of the traffic accident happens in the static region may be non-limiting, for example, 1, 2, 3, etc.
- the processing device 120 may determine that the rear-end collision happens in the static region.
- the processing device 120 may determine that both the collision between the vehicle and the preset fixture and the side flip of the vehicle happen in the static region.
- the processing device 120 may transmit a notification of the traffic accident result to a third party (e.g., a transportation surveillance department) .
- the third party may dispatch staff to process the traffic accident in real-time.
- the processing device 120 may transmit a notification of the traffic accident result to a terminal of a user (e.g., a driver) associated with the traffic accident. The notification may be used to confirm whether the traffic accident happens, whether the user needs a rescue, etc.
- the processing device 120 may transmit a notification of the traffic accident result to a monitor screen (e.g., implemented on the terminal 130) .
- the monitor screen may present (e.g., display, broadcast, etc. ) the notification of the traffic accident result in a form of text, image, video, voice, etc.
- the processing device 120 may store information and/or data (e.g., the traffic accident result) associated with the traffic accident detection in a storage device (e.g., the storage device 150, the ROM 230, the RAM 240, and/or the storage 390) disclosed elsewhere in the present disclosure.
- a storage device e.g., the storage device 150, the ROM 230, the RAM 240, and/or the storage 390
- FIG. 12 is a schematic diagram illustrating an exemplary first model according to some embodiments of the present disclosure.
- a first model may be configured to output a traffic accident result in a static region by processing the static region associated with a target scene in a target image.
- the traffic accident result may include a result indicating whether a traffic accident happens in the static region, a confidence value associated with the traffic accident, a type of the traffic accident, etc.
- the first model may include a first structure including a convolutional layer and a nonlinear layer, a second structure including a max-pooling layer, a convolutional layer, and a nonlinear layer, two linear layers, and a softmax layer.
- a count of the first structure may be 3.
- a count of the second structure may be 2.
- the first model may determine the confidence value associated with the traffic accident and encode the confidence value. If the confidence value exceeds a value threshold (e.g., 0.5) , the confidence value may be encoded as “1” ; if the confidence value is smaller than or equal to the value threshold, the confidence value may be encoded as “0” . As shown in FIG. 12, the first model may output traffic accident results (e.g., 0, 1, 1, 1, 1, 0) of six images, where “1” indicates that there is a traffic accident in an image; “0” indicates that there is no traffic accident in an image.
- traffic accident results e.g., 0, 1, 1, 1, 1, 0
- FIG. 13 is a block diagram illustrating an exemplary accident detection system according to some embodiments of the present disclosure.
- the accident detection system 1300 may be an example of the accident detection system 100 in FIG. 1.
- the accident detection system 1300 may include an image capture device 1310 (e.g., the image capture device 110 in FIG. 1) and a detection device 1320 (e.g., the processing device 120 in FIG. 1) .
- the image capture device 1310 may be pre-mounted in a region where a traffic condition needs to be monitored.
- the image capture device 1310 may acquire video data (e.g., a video) of the region in real-time and transmit the video data to the detection device 1320 automatically or under a request of the detection device 1320.
- the detection device 1320 may determine whether a traffic accident happens in the region by analyzing the video data, for example, based on processes 500, 1400-2100.
- FIG. 14 is a flowchart illustrating an exemplary process for determining a traffic accident result according to some embodiments of the present disclosure.
- the process 1400 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 1400.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 1400 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 14 and described below is not intended to be limiting.
- the processing device 120 may determine a target image (e.g., the target image in FIG. 5) from a video.
- the processing device 120 may extract a static region (e.g., the static region in FIG. 5) of the target image.
- the processing device 120 may determine a traffic accident result by performing traffic accident detection on the static region using a traffic accident detection model (also referred to as a first model) . More descriptions regarding the process 1400 may be found elsewhere in the present disclosure, for example, the process 500 and the descriptions thereof.
- a traffic accident detection model also referred to as a first model
- FIG. 15 is a flowchart illustrating an exemplary process for extracting a static region in a target image according to some embodiments of the present disclosure.
- the process 1500 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 1500.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 1500 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 15 and described below is not intended to be limiting.
- the processing device 120 may determine a target image by blurring a dynamic region of a first image (e.g., the first image in FIG. 5) using one or more images (also referred to as one or more second images, e.g., the one or more second images in FIG. 5) captured before the first image. More descriptions regarding operation 1510 may be found elsewhere in the present disclosure, for example, operation 520 and the descriptions thereof.
- the processing device 120 may extract a static region of the target image. More descriptions regarding operation 1520 may be found elsewhere in the present disclosure, for example, operation 530 and the descriptions thereof.
- FIG. 16 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure.
- the process 1600 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 1600.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 1600 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 16 and described below is not intended to be limiting.
- the processing device 120 may determine one or more bounding boxes corresponding to one or more static objects in a target image (e.g., the target image in FIG. 5) by performing object detection on the target image using a detection model (e.g., the detection model in FIG. 5) .
- a target image e.g., the target image in FIG. 5
- a detection model e.g., the detection model in FIG. 5
- the processing device 120 may determine a static region of the target image based on the one or more bounding boxes corresponding to one or more static objects. More descriptions regarding the process 1600 may be found elsewhere in the present disclosure, for example, operation 530 and the descriptions thereof.
- FIG. 17 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure.
- the process 1700 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 1700.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 1700 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 17 and described below is not intended to be limiting.
- the processing device 120 may determine a new bounding box (e.g., the combined bounding box in FIG. 17) by combining at least one of one or more bounding boxes (e.g., the one or more bounding boxes in FIG. 16) in a target image. At least one distance between the at least one bounding box may be smaller than or equal to a distance threshold.
- the processing device 120 may determine a static region of the target image based on the new bounding box. More descriptions regarding the process 1600 may be found elsewhere in the present disclosure, for example, operation 530 and the descriptions thereof.
- FIG. 18 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure.
- the process 1800 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 1800.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 1800 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 18 and described below is not intended to be limiting.
- the processing device 120 may determine one or more connected components corresponding to one or more static objects in a target image by performing object segmentation on the target image using a segmentation model.
- the processing device 120 may determine a static region of the target image based on the one or more connected components corresponding to the one or more static objects. More descriptions regarding the process 1800 may be found elsewhere in the present disclosure, for example, operation 530 and the descriptions thereof.
- FIG. 19 is a flowchart illustrating an exemplary process for determining a static region in a target image according to some embodiments of the present disclosure.
- the process 1900 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 1900.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 1900 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 19 and described below is not intended to be limiting.
- the processing device 120 may determine one or more bounding boxes corresponding to one or more connected components (e.g., the one or more connected components in FIG. 18) in a target image (e.g., the target image in FIG. 18) .
- the processing device 120 may determine a new bounding box (e.g., the combined bounding box in FIG. 5) by combining at least one of the one or more bounding boxes in the target image that satisfies a condition (also referred to as a first condition) .
- the condition may relate to at least one distance between the one or more bounding boxes, more descriptions of which may be found in operation 530.
- the processing device 120 may determine a static region of the target image based on the new bounding box. More descriptions regarding the process 1900 may be found elsewhere in the present disclosure, for example, operation 530 and the descriptions thereof.
- FIG. 20 is a flowchart illustrating an exemplary process for modifying a confidence value associated with a traffic accident result according to some embodiments of the present disclosure.
- the process 2000 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 2000.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 2000 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 20 and described below is not intended to be limiting.
- the processing device 120 may determine one or more 3D bounding boxes of one or more static objects by performing a 3D transformation on the one or more static objects.
- the one or more static objects may be partially overlapped.
- the processing device 120 may determine an overlapping region of the one or more 3D bounding boxes and determine a ratio between a volume of the overlapping region and a volume of one of the one or more 3D bounding boxes.
- the processing device 120 may modify a confidence value associated with a traffic accident result based on the ratio. More descriptions regarding the process 2000 may be found elsewhere in the present disclosure, for example, operation 540 and the descriptions thereof.
- FIG. 21 is a flowchart illustrating an exemplary process for modifying a confidence value associated with a traffic accident result according to some embodiments of the present disclosure.
- the process 2100 may be implemented as a set of instructions (e.g., an application) stored in the storage ROM 230 or RAM 240.
- the processor 220 and/or the modules in FIG. 4 may execute the set of instructions, and when executing the instructions, the processor 220 and/or the modules may be configured to perform the process 2100.
- the operations of the illustrated process presented below are intended to be illustrative. In some embodiments, the process 2100 may be accomplished with one or more additional operations not described and/or without one or more of the operations herein discussed. Additionally, the order in which the operations of the process illustrated in FIG. 21 and described below is not intended to be limiting.
- the processing device 120 may determine a weighted summation result by performing a weighted summation on a ratio (e.g., the ratio in FIG. 20) and a confidence value (e.g., the confidence value in FIG. 20) .
- a ratio e.g., the ratio in FIG. 20
- a confidence value e.g., the confidence value in FIG. 20
- the processing device 120 may modify the confidence value based on the weighted summation result. In some embodiments, the processing device 120 may designate a value corresponding to the weighted summation result as the modified confidence value. More descriptions regarding the process 2100 may be found elsewhere in the present disclosure, for example, operation 540 and the descriptions thereof.
- FIG. 22 is a block diagram illustrating an exemplary electronic device according to some embodiments of the present disclosure.
- the electronic device 2200 may include a processor 2210 and a storage 2220 operably coupled to the processor 2210.
- the storage 2220 may be configured to store a program instruction for implementing any traffic accident detection process (e.g., the process 500, 1400- 2100) described in the present disclosure.
- the processor 2210 may be configured to execute the program instruction stored by the storage 2220 to implement any traffic accident detection process (e.g., the process 500, 1400-2100) described in the present disclosure.
- the processor 2210 may also be referred to as a central processing unit (CPU) .
- the processor 2210 may be an integrated circuit chip with signal processing capability.
- the processor 2210 may also be a general-purpose processor, a digital signal processor (DSP) , a dedicated integrated circuit (ASIC) , a field programmable gate array (FPGA) , or other programmable logic devices, a discrete logic device, a transistor logic device, a discrete hardware component, or the like, or any combination thereof.
- DSP digital signal processor
- ASIC dedicated integrated circuit
- FPGA field programmable gate array
- the general-purpose processor may include a microprocessor or any general processor.
- FIG. 23 is a block diagram illustrating an exemplary computer readable storage medium according to some embodiments of the present disclosure.
- the computer readable storage medium 2300 may store a program instruction 2310 used to execute any traffic accident detection process (e.g., the process 500, 1400-2100) described in the present disclosure.
- the program instruction 2310 may form a program file and be stored in the computer readable storage medium 2300 in a form of a software product, such that a computer device (e.g., a personal computer, a server, a network device) or a processor may execute any traffic accident detection process (e.g., the process 500, 1400-2100) described in the present disclosure.
- the computer readable storage medium 2300 may include a medium that can store a program instruction, for example, a universal serial bus (USB) flash drive, a mobile hard disk, a read-only memory (ROM) , a random access memory (RAM) , a magnetic disk, or a terminal device such as an optical disk, a computer, a server, a mobile phone, a tablet computer, etc.
- a program instruction for example, a universal serial bus (USB) flash drive, a mobile hard disk, a read-only memory (ROM) , a random access memory (RAM) , a magnetic disk, or a terminal device such as an optical disk, a computer, a server, a mobile phone, a tablet computer, etc.
- USB universal serial bus
- ROM read-only memory
- RAM random access memory
- magnetic disk or a terminal device such as an optical disk, a computer, a server, a mobile phone, a tablet computer, etc.
- aspects of the present disclosure may be illustrated and described herein in any of a number of patentable classes or context including any new and useful process, machine, manufacture, or composition of matter, or any new and useful improvement thereof. Accordingly, aspects of the present disclosure may be implemented entirely hardware, entirely software (including firmware, resident software, micro-code, etc. ) or combining software and hardware implementation that may all generally be referred to herein as a “unit, ” “module, ” or “system. ” Furthermore, aspects of the present disclosure may take the form of a computer program product embodied in one or more computer readable media having computer readable program code embodied thereon.
- a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including electro-magnetic, optical, or the like, or any suitable combination thereof.
- a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that may communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
- Program code embodied on a computer readable signal medium may be transmitted using any appropriate medium, including wireless, wireline, optical fiber cable, RF, or the like, or any suitable combination of the foregoing.
- Computer program code for carrying out operations for aspects of the present disclosure may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB. NET, Python or the like, conventional procedural programming languages, such as the "C" programming language, Visual Basic, Fortran 2003, Perl, COBOL 2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages.
- the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN) , or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider) or in a cloud computing environment or offered as a service such as a Software as a Service (SaaS) .
- LAN local area network
- WAN wide area network
- SaaS Software as a Service
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Software Systems (AREA)
- Computational Linguistics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Traffic Control Systems (AREA)
- Image Analysis (AREA)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP21893665.6A EP4229547A4 (en) | 2020-11-20 | 2021-10-20 | SYSTEMS AND METHODS FOR DETECTING TRAFFIC ACCIDENTS |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011314812.5 | 2020-11-20 | ||
CN202011314812.5A CN112446316B (zh) | 2020-11-20 | 2020-11-20 | 事故检测方法、电子设备及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022105517A1 true WO2022105517A1 (en) | 2022-05-27 |
Family
ID=74737543
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/124913 WO2022105517A1 (en) | 2020-11-20 | 2021-10-20 | Systems and methods for detecting traffic accidents |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP4229547A4 (zh) |
CN (1) | CN112446316B (zh) |
WO (1) | WO2022105517A1 (zh) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112446316B (zh) * | 2020-11-20 | 2022-06-07 | 浙江大华技术股份有限公司 | 事故检测方法、电子设备及存储介质 |
CN113762219A (zh) * | 2021-11-03 | 2021-12-07 | 恒林家居股份有限公司 | 一种移动会议室内人物识别方法、系统和存储介质 |
CN115376311A (zh) * | 2022-07-08 | 2022-11-22 | 济南瑞源智能城市开发有限公司 | 一种针对隧道的车辆态势分析方法及设备 |
CN114913470B (zh) * | 2022-07-11 | 2022-10-28 | 浙江大华技术股份有限公司 | 一种事件检测方法及装置 |
CN115662112A (zh) * | 2022-09-30 | 2023-01-31 | 南京市德赛西威汽车电子有限公司 | 一种碰撞事件处理方法,处理系统,路侧单元和存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020196341A1 (en) * | 2001-06-21 | 2002-12-26 | Fujitsu Limited | Method and apparatus for processing pictures of mobile object |
CN101105892A (zh) * | 2007-07-30 | 2008-01-16 | 深圳市融合视讯科技有限公司 | 一种车辆交通事故的自动检测方法 |
CN102496000A (zh) * | 2011-11-14 | 2012-06-13 | 电子科技大学 | 一种城市交通事故检测方法 |
CN111369807A (zh) * | 2020-03-24 | 2020-07-03 | 北京百度网讯科技有限公司 | 一种交通事故的检测方法、装置、设备和介质 |
CN111800616A (zh) * | 2020-07-22 | 2020-10-20 | 多伦信息技术有限公司 | 一种交通事故智能检测系统 |
CN112446316A (zh) * | 2020-11-20 | 2021-03-05 | 浙江大华技术股份有限公司 | 事故检测方法、电子设备及存储介质 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103258432B (zh) * | 2013-04-19 | 2015-05-27 | 西安交通大学 | 基于视频的交通事故自动识别处理方法和系统 |
CN104183127B (zh) * | 2013-05-21 | 2017-02-22 | 北大方正集团有限公司 | 交通监控视频检测方法和装置 |
CN108510718A (zh) * | 2018-05-28 | 2018-09-07 | 深圳市零度智控科技有限公司 | 基于大数据的报警方法、装置、终端设备及可读存储介质 |
CN110659551A (zh) * | 2018-06-29 | 2020-01-07 | 比亚迪股份有限公司 | 运动状态的识别方法、装置及车辆 |
-
2020
- 2020-11-20 CN CN202011314812.5A patent/CN112446316B/zh active Active
-
2021
- 2021-10-20 EP EP21893665.6A patent/EP4229547A4/en active Pending
- 2021-10-20 WO PCT/CN2021/124913 patent/WO2022105517A1/en unknown
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020196341A1 (en) * | 2001-06-21 | 2002-12-26 | Fujitsu Limited | Method and apparatus for processing pictures of mobile object |
CN101105892A (zh) * | 2007-07-30 | 2008-01-16 | 深圳市融合视讯科技有限公司 | 一种车辆交通事故的自动检测方法 |
CN102496000A (zh) * | 2011-11-14 | 2012-06-13 | 电子科技大学 | 一种城市交通事故检测方法 |
CN111369807A (zh) * | 2020-03-24 | 2020-07-03 | 北京百度网讯科技有限公司 | 一种交通事故的检测方法、装置、设备和介质 |
CN111800616A (zh) * | 2020-07-22 | 2020-10-20 | 多伦信息技术有限公司 | 一种交通事故智能检测系统 |
CN112446316A (zh) * | 2020-11-20 | 2021-03-05 | 浙江大华技术股份有限公司 | 事故检测方法、电子设备及存储介质 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4229547A4 * |
Also Published As
Publication number | Publication date |
---|---|
EP4229547A1 (en) | 2023-08-23 |
EP4229547A4 (en) | 2024-04-24 |
CN112446316B (zh) | 2022-06-07 |
CN112446316A (zh) | 2021-03-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022105517A1 (en) | Systems and methods for detecting traffic accidents | |
US12087010B2 (en) | Systems and methods for lane detection | |
WO2018121013A1 (en) | Systems and methods for detecting objects in images | |
US11568548B2 (en) | Methods and systems for data visualization | |
US20220245792A1 (en) | Systems and methods for image quality detection | |
US20190138849A1 (en) | Rotation variant object detection in deep learning | |
CN109683699B (zh) | 基于深度学习实现增强现实的方法、装置及移动终端 | |
WO2019037129A1 (en) | METHODS AND SYSTEMS FOR DETECTING ENVIRONMENTAL INFORMATION OF A VEHICLE | |
US11538168B2 (en) | Incremental segmentation of point cloud | |
US20230341267A1 (en) | Systems and methods for temperature determination | |
WO2022105197A1 (en) | Systems and methods for image detection | |
US20210409761A1 (en) | Systems and methods for image coding | |
US20220221274A1 (en) | Positioning systems and methods | |
CN112927234A (zh) | 点云语义分割方法、装置、电子设备和可读存储介质 | |
WO2022247406A1 (en) | Systems and methods for determining key frame images of video data | |
WO2020113989A1 (en) | Method and system for image processing | |
US20230121534A1 (en) | Method and electronic device for 3d object detection using neural networks | |
CN116052026B (zh) | 一种无人机航拍图像目标检测方法、系统及存储介质 | |
US20220189297A1 (en) | Systems and methods for traffic monitoring | |
CN112001226A (zh) | 一种无人驾驶3d目标检测方法、装置及存储介质 | |
CN115147328A (zh) | 三维目标检测方法及装置 | |
US11657592B2 (en) | Systems and methods for object recognition | |
US20230087261A1 (en) | Three-dimensional target estimation using keypoints | |
CN113537017A (zh) | 基于级联型回归修正的光学遥感图像飞机检测方法和装置 | |
CN111881984A (zh) | 一种基于深度学习的目标检测方法和装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21893665 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021893665 Country of ref document: EP Effective date: 20230519 |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112023009790 Country of ref document: BR |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 112023009790 Country of ref document: BR Kind code of ref document: A2 Effective date: 20230519 |