WO2018058573A1 - Object detection method, object detection apparatus and electronic device - Google Patents

Object detection method, object detection apparatus and electronic device Download PDF

Info

Publication number
WO2018058573A1
WO2018058573A1 PCT/CN2016/101204 CN2016101204W WO2018058573A1 WO 2018058573 A1 WO2018058573 A1 WO 2018058573A1 CN 2016101204 W CN2016101204 W CN 2016101204W WO 2018058573 A1 WO2018058573 A1 WO 2018058573A1
Authority
WO
WIPO (PCT)
Prior art keywords
video image
image frame
interest
region
unit
Prior art date
Application number
PCT/CN2016/101204
Other languages
French (fr)
Chinese (zh)
Inventor
伍健荣
刘晓青
白向晖
谭志明
东明浩
Original Assignee
富士通株式会社
伍健荣
刘晓青
白向晖
谭志明
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 富士通株式会社, 伍健荣, 刘晓青, 白向晖, 谭志明 filed Critical 富士通株式会社
Priority to PCT/CN2016/101204 priority Critical patent/WO2018058573A1/en
Priority to CN201680087601.8A priority patent/CN109479118A/en
Publication of WO2018058573A1 publication Critical patent/WO2018058573A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present application relates to the field of information technology, and in particular, to a video image based object detecting method, an object detecting device, and an electronic device.
  • object detection can be performed on a video surveillance image, thereby identifying an object such as a specific vehicle, and further implementing functions such as object recognition, tracking, and control.
  • object detection can be performed on the entire image range of the video image frame, so that the blind area of detection can be avoided, but the range of detection needs to be large, and the data processing amount when detecting is compared. Big.
  • a Region of Interest may be preset in a video image frame, and object detection is performed for a region of interest of each video image frame, thereby reducing detection time.
  • the amount of data processing increases the detection speed.
  • the region of interest is preset, and the locations of the regions of interest in each video image frame are the same unless a new region of interest is re-set.
  • the object to be detected usually moves, and when it moves outside the region of interest, it is difficult to be detected, thereby causing a missed detection.
  • An embodiment of the present application provides an object detecting method, an object detecting apparatus, and an electronic device, which can extract a region of interest based on motion information of a video image frame, and perform object detection according to the extracted region of interest, Therefore, the accuracy of object detection can be improved and the detection speed can be improved.
  • an object detection apparatus For detecting a target object from a video image frame, the apparatus includes:
  • An extracting unit that extracts a region of interest from the video image frame based on motion information of a video image frame
  • a detecting unit that performs object detection in the video image frame according to the region of interest extracted by the extracting unit.
  • an object detecting method for detecting a target object from a video image frame, the method comprising:
  • Object detection is performed in the video image frame based on the extracted region of interest.
  • an electronic device comprising the object detecting device of the first aspect of the above embodiment.
  • the beneficial effects of the embodiments of the present application are that, according to the implementation of the present application, the accuracy of object detection can be improved, and the detection speed can be improved.
  • FIG. 1 is a schematic diagram of an object detecting device according to Embodiment 1 of the present application.
  • FIG. 2 is a schematic diagram of an extracting unit of Embodiment 1 of the present application.
  • FIG. 3 is a schematic diagram of a video image frame according to Embodiment 1 of the present application.
  • FIG. 4 is a schematic diagram of a binarized moving image corresponding to the video image frame of FIG. 3;
  • FIG. 5 is a schematic diagram of performing a connected domain segmentation process on the binarized moving image of FIG. 4 and generating a circumscribed rectangle;
  • FIG. 6 is a schematic diagram of merging connected domains according to Embodiment 1 of the present application.
  • FIG. 7 is another schematic diagram of merging connected domains according to Embodiment 1 of the present application.
  • FIG. 8 is a schematic diagram of a detecting unit of Embodiment 1 of the present application.
  • Embodiment 9 is a schematic diagram of combining detection results according to Embodiment 1 of the present application.
  • FIG. 11 is a schematic flow chart of an object detecting method according to Embodiment 2 of the present application.
  • FIG. 12 is a schematic diagram of a method for extracting a region of interest according to Embodiment 2 of the present application.
  • FIG. 13 is a schematic diagram of a method for performing object detection according to Embodiment 2 of the present application.
  • FIG. 14 is a schematic diagram showing the configuration of an electronic device according to Embodiment 3 of the present application.
  • Embodiment 1 of the present application provides an object detection device for detecting a target object from a video image frame.
  • the detecting device 100 includes an extracting unit 101 and a detecting unit 102.
  • the extracting unit 101 extracts a region of interest from the video image frame based on the motion information of the video image frame; the detecting unit 102 is in the video image frame according to the region of interest extracted by the extracting unit 101. Perform object detection.
  • the object detecting apparatus can extract the region of interest based on the motion information of the video image frame, and perform object detection based on the extracted region of interest, thereby being able to extract more accurately for each video image frame Corresponding regions of interest, thereby improving the accuracy of object detection and increasing the speed of detection.
  • the video image frame may be, for example, an image frame in a video captured by the surveillance camera.
  • the video image frame may also be from other devices, which is not limited in this embodiment.
  • the extracting unit 101 includes a motion detecting unit 201, a region dividing unit 202, and a generating unit 203.
  • the motion detecting unit 201 is configured to detect motion information in a video image frame; the region dividing unit 202 is configured to divide each moving object in the video image frame according to the motion information detected by the motion detecting unit 201.
  • the occupied area; the generating unit 203 generates at least one region of interest according to an area occupied by each moving object in the video image frame, the at least one region of interest covering an area where each moving object in the video image frame is located .
  • the motion detecting unit 201 may perform foreground detection on the video image frame to generate a binarized motion image of the video image frame, and according to the binarized moving image, the video image frame may be obtained.
  • the motion information may reflect motion information of the video image frame according to the first pixel in the binarized motion image, wherein the first pixel may be, for example, a white pixel.
  • FIG. 3 is a schematic diagram of a video image frame
  • FIG. 4 is a schematic diagram of a binarized moving image corresponding to the video image frame of FIG. 3, and the white pixels in the binarized moving image 400 of FIG. 4 can reflect the video image frame 300.
  • Sports information is a schematic diagram of a video image frame
  • the region dividing unit 202 may perform a connected domain segmentation process on the binarized moving image to obtain at least one connected domain of the pixel, where the at least one connected domain may correspond to each moving object in the video image frame.
  • Area For example, in the binarized moving image, each connected domain includes a plurality of first pixels, and within each connected domain, the first pixel is connected, and between the different connected domains, the first pixel is not connected, and thus different The connected areas are isolated from each other.
  • the region dividing unit 202 may further generate a circumscribed polygon of the connected domain for each connected domain in the binarized moving image, and the circumscribed polygon may be used to represent a contour of each connected domain, and the circumscribed polygon may be, for example, It is a rectangle or the like.
  • 5 is a schematic diagram of the connected domain segmentation process of the binarized moving image of FIG. 4 and the generation of the circumscribed rectangle. As shown in FIG. 5, each circumscribed rectangle 501 represents the contour of each connected domain, and The area enclosed by each circumscribed rectangle 501 corresponds to the area occupied by each moving object in the video image frame 300.
  • the area dividing unit 202 may also merge the connected domains whose distances are less than or equal to the first threshold as a new connected domain.
  • the first threshold may be a value greater than 0, and the distance between the connected domains may refer to the distance between the boundaries of the connected domains, or may refer to the geometric center or the centroid of each connected domain.
  • the area dividing unit 202 may also generate a circumscribed polygon by a new connected domain formed by combining at least two connected domains.
  • FIG. 6 is a schematic diagram of merging the connected domains.
  • the circumscribed rectangles of the two connected domains are 6011 and 6012, respectively, and the circumscribed rectangles 6011 and 6012 partially overlap.
  • the two connected domains are merged into the connected domain 6020, and the circumscribed rectangle of the connected domain 6020 is 6021, wherein the circumscribed rectangle is 6021, which may be a circumscribed rectangle of the circumscribed rectangles 6011 and 6012.
  • Figure 7 is another schematic diagram of the merging of connected domains.
  • the circumscribed rectangles of the four connected domains are 7011, 7012, 7013, and 7014, respectively, and the distance between the four circumscribed rectangles and the boundary of the adjacent circumscribed rectangle is smaller than the first Threshold.
  • the four connected domains are merged into the connected domain 7020, and the circumscribed rectangle of the connected domain 7020 is 7021.
  • the circumscribed rectangle 7021 may be a circumscribed rectangle 7011, 7012, 7013. And the circumscribed rectangle of 7014.
  • the distance between the circumscribed rectangle 7016 and the circumscribed rectangles 7011 7070 is far, for example, the distance is greater than the first threshold. Therefore, the connected domain corresponding to the circumscribed rectangle 7016 is not connected to the circumscribed rectangle 7011 ⁇ The connected domains corresponding to 7014 are merged.
  • the generating unit 203 is capable of generating at least one region of interest according to the distance between the regions occupied by the moving objects in the video image frame, whereby the regions closer to each other can be in the same interest. Within the scope covered by the area.
  • the generating unit 203 can be binarized according to the binarization.
  • the distance between the connected domains in the moving image to generate a region of interest for example, the generating unit 203 can make the distance A connected domain that is less than or equal to the second threshold is covered by the same region of interest.
  • the distance between the connected domain corresponding to the circumscribed rectangle 7016 and the connected domain 7020 is less than or equal to the second threshold. Therefore, the connected domain corresponding to the circumscribed rectangle 7016 and the connected domain 7020 are the same region of interest 703. Covered, wherein the boundary 7031 of the region of interest 703 is identified by a rectangular frame. Of course, the embodiment is not limited thereto, and the region of interest may be identified in other manners. For example, the boundary 7031 may be other polygonal frames.
  • the size of the boundary of the region of interest 703 may be larger than the size of the circumscribed polygon of each connected domain covered by it.
  • the size of the boundary 7031 of the region of interest 703 may be larger than the size of the circumscribed rectangle 7016 and the circumscribed rectangle 7021.
  • the size of the circumscribed rectangle, for example, the former can be 10% larger than the latter.
  • the generating unit 203 may use a corresponding region of the region of interest generated in the binarized moving image in the video image frame as the region of interest of the video image frame, whereby the extracting unit 101 can obtain the video image from the video image.
  • the region of interest is extracted from the frame.
  • Block 301 of FIG. 3 illustrates the boundaries of the region of interest extracted from the video image frame 300 by the extraction unit 101 in accordance with the present application.
  • the detecting unit 102 can perform object detection in the video image frame based on the region of interest extracted by the extracting unit 101.
  • FIG. 8 is a schematic diagram of the detecting unit 102. As shown in FIG. 8, the detecting unit 102 may include a determining unit 801 and an object detecting unit 802.
  • the determining unit 801 is configured to determine whether the number of regions of interest in the video image frame is less than or equal to a third threshold, and whether the area of the region of interest is less than or equal to a fourth threshold; the object detecting unit 802 determines The result of the determination by unit 801 is object detection in the region of interest of the video image frame or the entire image range of the video image frame.
  • the object detecting unit 802 performs object detection in each region of interest of the video image frame, whereby fast object detection can be performed.
  • the object detecting unit 802 does not perform object detection on the video image frame.
  • the determining unit 801 determines that the number of regions of interest in the video image frame is large At a third threshold, or the sum of the areas of the region of interest in the video image frame is greater than a fourth threshold, then in the video image frame, the object detecting unit 802 performs object detection in the entire image range of the video image frame, thereby Can prevent missed inspections.
  • the specific method for the object detection unit 802 to perform the object detection may refer to the prior art, and is not described in this embodiment.
  • a specific video image frame in the video may be used as a key frame, and other video image frames in the video may be used as a normal frame, wherein a video separated by a predetermined time may be used.
  • the image frame or the video image frame of a predetermined number of frames is used as a key frame.
  • other methods may be used to set the key frame.
  • the determining unit 801 can determine whether the video image frame is a normal frame, and for the normal frame, further determining, according to the determination result of the determining unit 801, performing object detection in the region of interest of the normal frame or the entire image range of the normal frame.
  • the determination unit 801 can be used for further determination, and the object detection can be performed directly in the entire image range of the key frame. Thereby, it is possible to prevent missed detection by performing object detection in the entire image range on the key frame.
  • the detecting unit 102 may further have a merging unit 803.
  • the merging unit 803 may detect the detection result in the region of interest of the current video image frame and the video image frame before the current video image frame.
  • the detection result is merged, and the combined detection result may include, for example, a detection result of the region of interest of the current video image frame, and detection of the previous video image frame outside the region of interest of the current video image frame. result.
  • FIG. 9 is a schematic diagram of merging detection results
  • 901 is a video image frame before the current video image frame 902
  • 9011, 9012 are object objects detected in the video image frame 901
  • the current video image frame 902 is sensed.
  • the region of interest is 9021
  • the target object 9022 is detected in the region of interest 9021
  • the detection result of the current video image frame 902 is combined with the detection result of the previous video image frame 901 to obtain a combined detection result 903, in the merged
  • the detection result 903 includes: the object object 9022 detected in the region of interest 9021 in the current video image frame 902, and the region other than the region of interest 9021 of the current video image frame 902, detected in the previous video image frame 901 Object object 9012.
  • Step 1001 The determining unit 801 determines whether the current video image frame is a normal frame, and if yes, proceeds to step 1002, and if no, proceeds to step 1005.
  • Step 1002 The determining unit 801 determines whether the number of regions of interest in the current video image frame is less than or equal to a third threshold, and if yes, proceeds to step 1003, and if no, proceeds to step 1005.
  • Step 1003 The determining unit 801 determines whether the total area of the region of interest in the current video image frame is less than or equal to the fourth threshold. If yes, proceed to step 1004. If no, proceed to step 1005.
  • Step 1004 The object detecting unit 802 performs object detection in the region of interest of the video image frame.
  • Step 1005 The object detecting unit 802 performs object detection in the entire image range of the video image frame.
  • Step 1006 The merging unit 803 combines the detection result of the region of interest of the current video image frame with the detection result of the previous video image frame.
  • the object detecting apparatus can extract the region of interest based on the motion information of the video image frame, and perform object detection based on the extracted region of interest, thereby being able to extract more accurately for each video image frame Corresponding regions of interest, thereby improving the accuracy of object detection and increasing the speed of detection.
  • the embodiment of the present application further provides an object detecting method for detecting a target object from a video image frame, corresponding to the object detecting device of Embodiment 1.
  • FIG. 11 is a schematic flowchart of the object detecting method in the second embodiment. As shown in FIG. 11, the detecting method may include:
  • Step 1101 extracting a region of interest from the video image frame based on motion information of a video image frame
  • Step 1102 Perform object detection in the video image frame according to the extracted region of interest.
  • FIG. 12 is a schematic diagram of a method for extracting a region of interest according to the second embodiment. As shown in FIG. 12, the method includes:
  • Step 1201 Detect motion information in the video image frame.
  • Step 1202 According to the detected motion information, divide an area occupied by each moving object in the video image frame;
  • Step 1203 Generate at least one region of interest according to an area occupied by each moving object in the video image frame, where the at least one region of interest covers an area where each moving object in the video image frame is located.
  • binarization of the video image frame may be generated based on foreground detection.
  • a motion image to obtain the motion information of the video image frame.
  • the connected domain segmentation process may be performed on the binarized moving image to obtain at least one connected domain of the pixel, where the at least one connected domain corresponds to each moving object in the video image frame. Occupied area.
  • an circumscribed polygon of each of the connected domains may also be generated.
  • the connected domains whose distances from each other are less than or equal to the first threshold may also be merged into one new connected domain.
  • the at least one region of interest may be generated according to the distance of the region.
  • FIG. 13 is a schematic diagram of a method for performing object detection in the video image frame according to the extracted region of interest according to the second embodiment. As shown in FIG. 13, the method includes:
  • Step 1301 Determine whether the number of the regions of interest in the video image frame is less than or equal to a third threshold, and whether an area of the region of interest is less than or equal to a fourth threshold;
  • Step 1302 Perform object detection in the region of interest of the video image frame or the entire image range of the video image frame according to the result of the determining.
  • the method further includes:
  • Step 1303 Combine the detection result in the region of interest of the current video image frame and the detection result of the video image frame before the current video image frame in the case of performing object detection on the region of interest of the current video image frame.
  • the object detecting method can extract the region of interest based on the motion information of the video image frame, and perform object detection based on the extracted region of interest, thereby being able to extract more accurately for each video image frame. Corresponding regions of interest, thereby improving the accuracy of object detection and increasing the speed of detection.
  • Embodiment 3 of the present application provides an electronic device including the object detecting device as described in Embodiment 1.
  • FIG. 14 is a schematic diagram showing the configuration of an electronic device according to Embodiment 3 of the present application.
  • the electronic device 1400 can include a central processing unit (CPU) 1401 and a memory 1402; the memory 1402 is coupled to the center.
  • the memory 1402 can store various data; in addition, a program for performing object detection is stored, and the program is executed under the control of the central processing unit 1401.
  • the functionality in the object detection device can be integrated into the central processor 1401.
  • the central processing unit 1401 can be configured to:
  • Object detection is performed in the video image frame based on the extracted region of interest.
  • the central processor 1401 can also be configured to:
  • the central processor 1401 can also be configured to:
  • a binarized motion image of the video image frame is generated based on foreground detection, thereby obtaining the motion information of the video image frame.
  • the central processor 1401 can also be configured to:
  • Connected domain segmentation processing is performed on the binarized moving image to obtain at least one connected domain of the pixel, the at least one connected domain corresponding to an area occupied by each moving object in the video image frame.
  • the central processor 1401 can also be configured to:
  • the central processor 1401 can also be configured to:
  • the connected domains that are less than or equal to the first threshold are merged into a new connected domain.
  • the central processor 1401 can also be configured to:
  • the at least one region of interest is generated based on the distance of the region.
  • the central processor 1401 can also be configured to:
  • object detection is performed in the region of interest of the video image frame or in the entire image range of the video image frame.
  • the central processor 1401 can also be configured to:
  • the detection results in the region of interest of the current video image frame and the detection results of the video image frames preceding the current video image frame are combined.
  • the electronic device 1400 may further include: an input and output unit 1403, a display unit 1404, and the like; wherein the functions of the above components are similar to those of the prior art, and details are not described herein again. It should be noted that the electronic device 1400 does not necessarily have to include all the components shown in FIG. 14; in addition, the electronic device 1400 may further include components not shown in FIG. 14, and reference may be made to the prior art.
  • the embodiment of the present application further provides a computer readable program, wherein the program causes the object detecting device or the electronic device to perform the object detection described in Embodiment 2 when the program is executed in an object detecting device or an electronic device method.
  • the embodiment of the present application further provides a storage medium storing a computer readable program, wherein the storage medium stores the computer readable program, wherein the computer readable program causes the object detecting device or the electronic device to perform the embodiment 2 Object detection method.
  • the object detecting apparatus described in connection with the embodiments of the present invention may be directly embodied as hardware, a software module executed by a processor, or a combination of both.
  • one or more of the functional blocks shown in Figures 1, 2, and 8 and/or one or more combinations of functional blocks may correspond to individual software modules of a computer program flow, or to individual hardware.
  • These software modules may correspond to the respective steps shown in Embodiment 2, respectively.
  • These hardware modules can be implemented, for example, by curing these software modules using a Field Programmable Gate Array (FPGA).
  • FPGA Field Programmable Gate Array
  • the software module can reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, removable disk, CD-ROM, or any other form of storage medium known in the art.
  • a storage medium can be coupled to the processor to enable the processor to read information from, and write information to, the storage medium; or the storage medium can be an integral part of the processor.
  • the processor and the storage medium can be located in an ASIC.
  • the software module can be stored in the memory of the mobile terminal or in a memory card that can be inserted into the mobile terminal.
  • the software module can be stored in the MEGA-SIM card or a large-capacity flash memory device.
  • One or more of the functional block diagrams described with respect to Figures 1, 2, 8 and/or one or more groups of functional block diagrams A general purpose processor, digital signal processor (DSP), application specific integrated circuit (ASIC), field programmable gate array (FPGA) or other programmable logic device, discrete door for performing the functions described herein can be implemented. Or transistor logic device, discrete hardware component, or any suitable combination thereof.
  • One or more of the functional blocks described with respect to Figures 1-3 and/or one or more combinations of functional blocks may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, multiple microprocessors One or more microprocessors in conjunction with DSP communication or any other such configuration.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

Provided in the embodiments of the present application are an object detection method, an object detection apparatus and an electronic device, for detecting an object from a video image frame. The object detection apparatus comprises: an extraction unit for extracting, on the basis of movement information of a video image frame, a region of interest from the video image frame; and a detection unit for performing, according to the region of interest extracted by the extraction unit, object detection on the video image frame. According to the present application, the accuracy and speed of object detection are improved.

Description

对象检测方法、对象检测装置以及电子设备Object detection method, object detection device, and electronic device 技术领域Technical field
本申请涉及信息技术领域,特别涉及一种基于视频图像的对象检测方法、对象检测装置以及电子设备。The present application relates to the field of information technology, and in particular, to a video image based object detecting method, an object detecting device, and an electronic device.
背景技术Background technique
随着信息技术的发展,基于图像的对象检测技术被越来越广泛地应用。例如,在交通监控领域,可以针对视频监控图像进行对象检测,从而识别出特定的车辆等对象,并进而实现对象的识别、跟踪、控制等功能。With the development of information technology, image-based object detection technology is more and more widely used. For example, in the field of traffic monitoring, object detection can be performed on a video surveillance image, thereby identifying an object such as a specific vehicle, and further implementing functions such as object recognition, tracking, and control.
现有的基于视频图像的对象检测技术中,可以对视频图像帧的整个图像范围进行对象检测,这样,能够避免检测的盲区,但是,需要检测的范围较大,进行检测时的数据处理量较大。In the existing video image-based object detection technology, object detection can be performed on the entire image range of the video image frame, so that the blind area of detection can be avoided, but the range of detection needs to be large, and the data processing amount when detecting is compared. Big.
在现有技术中,还可以在视频图像帧中预先设定感兴趣区域(Region of Interest,ROI),并且,针对每个视频图像帧的感兴趣区域进行对象检测,这样,能够减少进行检测时的数据处理量,提高检测速度。In the prior art, a Region of Interest (ROI) may be preset in a video image frame, and object detection is performed for a region of interest of each video image frame, thereby reducing detection time. The amount of data processing increases the detection speed.
应该注意,上面对技术背景的介绍只是为了方便对本申请的技术方案进行清楚、完整的说明,并方便本领域技术人员的理解而阐述的。不能仅仅因为这些方案在本申请的背景技术部分进行了阐述而认为上述技术方案为本领域技术人员所公知。It should be noted that the above description of the technical background is only for the purpose of facilitating a clear and complete description of the technical solutions of the present application, and is convenient for understanding by those skilled in the art. The above technical solutions are not considered to be well known to those skilled in the art simply because these aspects are set forth in the background section of this application.
申请内容Application content
本申请的发明人发现,在现有技术中,感兴趣区域是预先设定的,并且,各视频图像帧中的感兴趣区域的位置都一样,除非重新设定新的感兴趣区域。但是,在应用对象检测的场景中,需要被检测出的对象通常会运动,当其运动到感兴趣区域之外时,就难以被检测到,从而造成漏检。The inventors of the present application found that in the prior art, the region of interest is preset, and the locations of the regions of interest in each video image frame are the same unless a new region of interest is re-set. However, in the scene in which the object is detected, the object to be detected usually moves, and when it moves outside the region of interest, it is difficult to be detected, thereby causing a missed detection.
本申请实施例提供一种对象检测方法、对象检测装置以及电子设备,该对象检测装置能够基于视频图像帧的运动信息来提取感兴趣区域,并根据提取出的感兴趣区域来进行对象检测,由此,能够提高对象检测的准确性,并提高检测速度。An embodiment of the present application provides an object detecting method, an object detecting apparatus, and an electronic device, which can extract a region of interest based on motion information of a video image frame, and perform object detection according to the extracted region of interest, Therefore, the accuracy of object detection can be improved and the detection speed can be improved.
根据本申请实施例的第一方面,提供了一种对象检测(object detection)装置, 用于从视频图像帧中检测出对象物体,该装置包括:According to a first aspect of embodiments of the present application, an object detection apparatus is provided, For detecting a target object from a video image frame, the apparatus includes:
提取单元,其基于视频图像帧的运动信息,从所述视频图像帧中提取出感兴趣区域;以及An extracting unit that extracts a region of interest from the video image frame based on motion information of a video image frame;
检测单元,其根据所述提取单元所提取出的感兴趣区域,在所述视频图像帧中进行对象检测。And a detecting unit that performs object detection in the video image frame according to the region of interest extracted by the extracting unit.
根据本申请实施例的第二方面,提供了一种对象检测方法,用于从视频图像帧中检测出对象物体,该方法包括:According to a second aspect of the embodiments of the present application, an object detecting method is provided for detecting a target object from a video image frame, the method comprising:
基于视频图像帧的运动信息,从所述视频图像帧中提取出感兴趣区域;以及Extracting a region of interest from the video image frame based on motion information of the video image frame;
根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测。Object detection is performed in the video image frame based on the extracted region of interest.
根据本申请实施例的第三方面,提供一种电子设备,包括上述实施例第一方面所述的对象检测装置。According to a third aspect of the embodiments of the present application, there is provided an electronic device comprising the object detecting device of the first aspect of the above embodiment.
本申请实施例的有益效果在于:根据本申请实施里,能够提高对象检测的准确性,并提高检测速度。The beneficial effects of the embodiments of the present application are that, according to the implementation of the present application, the accuracy of object detection can be improved, and the detection speed can be improved.
参照后文的说明和附图,详细公开了本申请的特定实施方式,指明了本申请的原理可以被采用的方式。应该理解,本申请的实施方式在范围上并不因而受到限制。在所附权利要求的精神和条款的范围内,本申请的实施方式包括许多改变、修改和等同。Specific embodiments of the present application are disclosed in detail with reference to the following description and accompanying drawings, in which <RTIgt; It should be understood that the embodiments of the present application are not limited in scope. The embodiments of the present application include many variations, modifications, and equivalents within the scope of the appended claims.
针对一种实施方式描述和/或示出的特征可以以相同或类似的方式在一个或更多个其它实施方式中使用,与其它实施方式中的特征相组合,或替代其它实施方式中的特征。Features described and/or illustrated with respect to one embodiment may be used in one or more other embodiments in the same or similar manner, in combination with, or in place of, features in other embodiments. .
应该强调,术语“包括/包含”在本文使用时指特征、整件、步骤或组件的存在,但并不排除一个或更多个其它特征、整件、步骤或组件的存在或附加。It should be emphasized that the term "comprising" or "comprises" or "comprising" or "comprising" or "comprising" or "comprising" or "comprises"
附图说明DRAWINGS
所包括的附图用来提供对本申请实施例的进一步的理解,其构成了说明书的一部分,用于例示本申请的实施方式,并与文字描述一起来阐释本申请的原理。显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。在附图中:The drawings are included to provide a further understanding of the embodiments of the present application, and are intended to illustrate the embodiments of the present application Obviously, the drawings in the following description are only some of the embodiments of the present application, and those skilled in the art can obtain other drawings according to the drawings without any inventive labor. In the drawing:
图1是本申请实施例1的对象检测装置的一个示意图; 1 is a schematic diagram of an object detecting device according to Embodiment 1 of the present application;
图2是本申请实施例1的提取单元的一个示意图;2 is a schematic diagram of an extracting unit of Embodiment 1 of the present application;
图3是本申请实施例1的视频图像帧的一个示意图;3 is a schematic diagram of a video image frame according to Embodiment 1 of the present application;
图4是图3的视频图像帧所对应的二值化运动图像的一个示意图;4 is a schematic diagram of a binarized moving image corresponding to the video image frame of FIG. 3;
图5是对图4的二值化运动图像进行连通域分割处理并生成外接矩形后的一个示意图;5 is a schematic diagram of performing a connected domain segmentation process on the binarized moving image of FIG. 4 and generating a circumscribed rectangle;
图6是本申请实施例1的对连通域进行合并的一个示意图;6 is a schematic diagram of merging connected domains according to Embodiment 1 of the present application;
图7是本申请实施例1的对连通域进行合并的另一个示意图;FIG. 7 is another schematic diagram of merging connected domains according to Embodiment 1 of the present application; FIG.
图8是本申请实施例1的检测单元的一个示意图;8 is a schematic diagram of a detecting unit of Embodiment 1 of the present application;
图9是本申请实施例1的对检测结果进行合并的一个示意图;9 is a schematic diagram of combining detection results according to Embodiment 1 of the present application;
图10是本申请实施例1的检测单元的一个工作流程图;10 is a working flow chart of the detecting unit of Embodiment 1 of the present application;
图11是本申请实施例2的对象检测方法的一个流程示意图;11 is a schematic flow chart of an object detecting method according to Embodiment 2 of the present application;
图12是本申请实施例2的提取出感兴趣区域的方法的一个示意图;FIG. 12 is a schematic diagram of a method for extracting a region of interest according to Embodiment 2 of the present application; FIG.
图13是本申请实施例2的进行对象检测的方法的一个示意图;FIG. 13 is a schematic diagram of a method for performing object detection according to Embodiment 2 of the present application; FIG.
图14是本申请实施例3的电子设备的一个构成示意图。FIG. 14 is a schematic diagram showing the configuration of an electronic device according to Embodiment 3 of the present application.
具体实施方式detailed description
参照附图,通过下面的说明书,本申请的前述以及其它特征将变得明显。在说明书和附图中,具体公开了本申请的特定实施方式,其表明了其中可以采用本申请的原则的部分实施方式,应了解的是,本申请不限于所描述的实施方式,相反,本申请包括落入所附权利要求的范围内的全部修改、变型以及等同物。下面结合附图对本申请的各种实施方式进行说明。这些实施方式只是示例性的,不是对本申请的限制。The foregoing and other features of the present application will be apparent from the description, The specific embodiments of the present application are specifically disclosed in the specification and the drawings, which illustrate a part of the embodiments in which the principles of the present application may be employed, it being understood that the present application is not limited to the described embodiments, but instead The application includes all modifications, variations and equivalents falling within the scope of the appended claims. Various embodiments of the present application will be described below with reference to the accompanying drawings. These embodiments are merely exemplary and are not limiting of the application.
实施例1Example 1
本申请实施例1提供一种对象检测(object detection)装置,用于从视频图像帧中检测出对象物体。Embodiment 1 of the present application provides an object detection device for detecting a target object from a video image frame.
图1是本实施例1的对象检测装置的一个示意图,如图1所示,该检测装置100包括提取单元101和检测单元102。1 is a schematic diagram of an object detecting device of the first embodiment. As shown in FIG. 1, the detecting device 100 includes an extracting unit 101 and a detecting unit 102.
其中,提取单元101基于视频图像帧的运动信息,从该视频图像帧中提取出感兴趣区域;检测单元102根据提取单元101所提取出的感兴趣区域,在该视频图像帧中 进行对象检测。The extracting unit 101 extracts a region of interest from the video image frame based on the motion information of the video image frame; the detecting unit 102 is in the video image frame according to the region of interest extracted by the extracting unit 101. Perform object detection.
根据本实施例,对象检测装置能够基于视频图像帧的运动信息来提取感兴趣区域,并根据提取出的感兴趣区域来进行对象检测,由此,能够针对每一个视频图像帧来更加准确地提取相应的感兴趣区域,从而提高对象检测的准确性,并提高检测速度。According to the present embodiment, the object detecting apparatus can extract the region of interest based on the motion information of the video image frame, and perform object detection based on the extracted region of interest, thereby being able to extract more accurately for each video image frame Corresponding regions of interest, thereby improving the accuracy of object detection and increasing the speed of detection.
在本实施例中,视频图像帧例如可以是监控摄像机所拍摄的视频中的图像帧,当然,该视频图像帧也可以来自于其他的装置,本实施例并不做限定。In this embodiment, the video image frame may be, for example, an image frame in a video captured by the surveillance camera. Of course, the video image frame may also be from other devices, which is not limited in this embodiment.
图2是本实施例的提取单元101的一个示意图,如图2所示,提取单元101包括运动检测单元201、区域划分单元202和生成单元203。2 is a schematic diagram of the extracting unit 101 of the present embodiment. As shown in FIG. 2, the extracting unit 101 includes a motion detecting unit 201, a region dividing unit 202, and a generating unit 203.
在本实施例中,运动检测单元201用于检测视频图像帧中的运动信息;区域划分单元202用于根据运动检测单元201所检测到的运动信息,划分出该视频图像帧中的各运动物体所占据的区域;生成单元203根据该视频图像帧中的各运动物体所占据的区域,生成至少一个感兴趣区域,该至少一个感兴趣区域覆盖该视频图像帧中的各运动物体所处的区域。In this embodiment, the motion detecting unit 201 is configured to detect motion information in a video image frame; the region dividing unit 202 is configured to divide each moving object in the video image frame according to the motion information detected by the motion detecting unit 201. The occupied area; the generating unit 203 generates at least one region of interest according to an area occupied by each moving object in the video image frame, the at least one region of interest covering an area where each moving object in the video image frame is located .
在本实施例中,运动检测单元201可以对视频图像帧进行前景检测,以生成该视频图像帧的二值化运动图像(motion image),根据该二值化运动图像,可以获得该视频图像帧的运动信息,例如,根据该二值化运动图像中的第一像素可以反映该视频图像帧的运动信息,其中,该第一像素例如可以是白色像素。In this embodiment, the motion detecting unit 201 may perform foreground detection on the video image frame to generate a binarized motion image of the video image frame, and according to the binarized moving image, the video image frame may be obtained. The motion information, for example, may reflect motion information of the video image frame according to the first pixel in the binarized motion image, wherein the first pixel may be, for example, a white pixel.
图3是视频图像帧的一个示意图,图4是图3的视频图像帧所对应的二值化运动图像的一个示意图,图4的二值化运动图像400中的白色像素能够反映视频图像帧300的运动信息。3 is a schematic diagram of a video image frame, FIG. 4 is a schematic diagram of a binarized moving image corresponding to the video image frame of FIG. 3, and the white pixels in the binarized moving image 400 of FIG. 4 can reflect the video image frame 300. Sports information.
在本实施例中,区域划分单元202可以对二值化运动图像进行连通域分割处理,以得到像素的至少一个连通域,该至少一个连通域可以对应于视频图像帧中的各运动物体所占据的区域。例如,在二值化运动图像中,各连通域中都包括多个第一像素,在各连通域的内部,第一像素连接,在不同的连通域之间,第一像素不连接,因而不同的连通区域彼此隔离。In this embodiment, the region dividing unit 202 may perform a connected domain segmentation process on the binarized moving image to obtain at least one connected domain of the pixel, where the at least one connected domain may correspond to each moving object in the video image frame. Area. For example, in the binarized moving image, each connected domain includes a plurality of first pixels, and within each connected domain, the first pixel is connected, and between the different connected domains, the first pixel is not connected, and thus different The connected areas are isolated from each other.
在本实施例中,区域划分单元202还可以为二值化运动图像中的每个连通域生成该连通域的外接多边形,该外接多边形能够用来表示各连通域的轮廓,该外接多边形例如可以是矩形等。图5是对图4的二值化运动图像进行连通域分割处理并生成外接矩形后的一个示意图,如图5所示,各外接矩形501分别代表各连通域的轮廓,并且, 各外接矩形501所围成的区域对应于视频图像帧300中各运动物体所占据的区域。In this embodiment, the region dividing unit 202 may further generate a circumscribed polygon of the connected domain for each connected domain in the binarized moving image, and the circumscribed polygon may be used to represent a contour of each connected domain, and the circumscribed polygon may be, for example, It is a rectangle or the like. 5 is a schematic diagram of the connected domain segmentation process of the binarized moving image of FIG. 4 and the generation of the circumscribed rectangle. As shown in FIG. 5, each circumscribed rectangle 501 represents the contour of each connected domain, and The area enclosed by each circumscribed rectangle 501 corresponds to the area occupied by each moving object in the video image frame 300.
在本实施例中,区域划分单元202还可以将彼此距离小于或等于第一阈值的连通域进行合并,以作为一个新的连通域。In this embodiment, the area dividing unit 202 may also merge the connected domains whose distances are less than or equal to the first threshold as a new connected domain.
在本实施例中,该第一阈值可以是大于0的值,连通域之间的距离可以是指各连通域的边界之间的距离,也可以是指各连通域的几何中心或质心之间的距离等;并且,在各连通域都具有外界多边形的情况下,连通域之间的距离可以是指各外接多边形连的边界之间的距离,也可以是指各外接多边形的几何中心或质心之间的距离等,其中,如果两个连通域外接多边形彼此有部分重叠,那么可以认为该两个连通域之间的距离为负值,小于该第一阈值。In this embodiment, the first threshold may be a value greater than 0, and the distance between the connected domains may refer to the distance between the boundaries of the connected domains, or may refer to the geometric center or the centroid of each connected domain. The distance and the like; and, in the case where each connected domain has an outer polygon, the distance between the connected domains may refer to the distance between the boundaries of the contiguous polygons, or may refer to the geometric center or centroid of each circumscribed polygon. The distance between the two, etc., wherein if the two connected domains circumscribed polygons partially overlap each other, the distance between the two connected domains can be considered to be a negative value, which is smaller than the first threshold.
在本实施例中,区域划分单元202也可以为通过将至少两个连通域进行合并而形成的新的连通域生成外接多边形。In this embodiment, the area dividing unit 202 may also generate a circumscribed polygon by a new connected domain formed by combining at least two connected domains.
图6是对连通域进行合并的一个示意图,如图6所示,合并前的图601中,两个连通域的外接矩形分别为6011和6012,并且,外接矩形6011和6012部分重叠。合并后的图602中,两个连通域被合并为连通域6020,并且,连通域6020的外接矩形为6021,其中,外接矩形为6021可以是外接矩形6011和6012的外接矩形。FIG. 6 is a schematic diagram of merging the connected domains. As shown in FIG. 6, in the pre-merge diagram 601, the circumscribed rectangles of the two connected domains are 6011 and 6012, respectively, and the circumscribed rectangles 6011 and 6012 partially overlap. In the merged diagram 602, the two connected domains are merged into the connected domain 6020, and the circumscribed rectangle of the connected domain 6020 is 6021, wherein the circumscribed rectangle is 6021, which may be a circumscribed rectangle of the circumscribed rectangles 6011 and 6012.
图7是对连通域进行合并的另一个示意图。如图7所示,合并前的图701中,四个连通域的外接矩形分别为7011、7012、7013和7014,并且,这四个外接矩形与相邻外接矩形的边界的距离小于该第一阈值。在区域划分单元202进行合并处理后的图702中,四个连通域被合并为连通域7020,并且,连通域7020的外接矩形为7021,其中,外接矩形7021可以是外接矩形7011、7012、7013和7014的外接矩形。Figure 7 is another schematic diagram of the merging of connected domains. As shown in FIG. 7, in the pre-merge diagram 701, the circumscribed rectangles of the four connected domains are 7011, 7012, 7013, and 7014, respectively, and the distance between the four circumscribed rectangles and the boundary of the adjacent circumscribed rectangle is smaller than the first Threshold. In the diagram 702 after the merging process is performed by the zoning unit 202, the four connected domains are merged into the connected domain 7020, and the circumscribed rectangle of the connected domain 7020 is 7021. The circumscribed rectangle 7021 may be a circumscribed rectangle 7011, 7012, 7013. And the circumscribed rectangle of 7014.
此外,如图7所示,外接矩形7016与外接矩形7011~7014的距离都较远,例如,该距离大于该第一阈值,因此,外接矩形7016所对应的连通域就不与外接矩形7011~7014所对应的连通域进行合并。In addition, as shown in FIG. 7 , the distance between the circumscribed rectangle 7016 and the circumscribed rectangles 7011 7070 is far, for example, the distance is greater than the first threshold. Therefore, the connected domain corresponding to the circumscribed rectangle 7016 is not connected to the circumscribed rectangle 7011 ~ The connected domains corresponding to 7014 are merged.
在本实施例中,生成单元203能够根据视频图像帧中的各运动物体所占据的区域之间的距离,生成至少一个感兴趣区域,由此,距离较近的区域能够处于到同一个感兴趣区域所覆盖的范围内。In the present embodiment, the generating unit 203 is capable of generating at least one region of interest according to the distance between the regions occupied by the moving objects in the video image frame, whereby the regions closer to each other can be in the same interest. Within the scope covered by the area.
在本实施例中,由于视频图像帧中的各运动物体所占据的区域之间的距离可以与二值化运动图像中的连通域之间的距离对应,因此,生成单元203可以根据二值化运动图像中的连通域之间的距离,来生成感兴趣区域,例如,生成单元203可以使距离 小于或等于第二阈值的连通域被同一个感兴趣区域所覆盖。In this embodiment, since the distance between the regions occupied by the moving objects in the video image frame can correspond to the distance between the connected domains in the binarized moving image, the generating unit 203 can be binarized according to the binarization. The distance between the connected domains in the moving image to generate a region of interest, for example, the generating unit 203 can make the distance A connected domain that is less than or equal to the second threshold is covered by the same region of interest.
如图7所示,外接矩形7016所对应的连通域与连通域7020之间的距离小于或等于第二阈值,因此,外接矩形7016所对应的连通域与连通域7020被同一个感兴趣区域703所覆盖,其中,该感兴趣区域703的边界7031用矩形框来标识。当然,本实施例不限于此,也可以用其他的方式来标识感兴趣区域,例如边界7031可以是其它的多边形框。As shown in FIG. 7, the distance between the connected domain corresponding to the circumscribed rectangle 7016 and the connected domain 7020 is less than or equal to the second threshold. Therefore, the connected domain corresponding to the circumscribed rectangle 7016 and the connected domain 7020 are the same region of interest 703. Covered, wherein the boundary 7031 of the region of interest 703 is identified by a rectangular frame. Of course, the embodiment is not limited thereto, and the region of interest may be identified in other manners. For example, the boundary 7031 may be other polygonal frames.
在图7中,该感兴趣区域703的边界的尺寸可以大于其所覆盖的各连通域的外接多边形的尺寸,例如,感兴趣区域703的边界7031的尺寸可以大于外接矩形7016和外接矩形7021的外接矩形的尺寸,比如,前者可以比后者大10%。In FIG. 7, the size of the boundary of the region of interest 703 may be larger than the size of the circumscribed polygon of each connected domain covered by it. For example, the size of the boundary 7031 of the region of interest 703 may be larger than the size of the circumscribed rectangle 7016 and the circumscribed rectangle 7021. The size of the circumscribed rectangle, for example, the former can be 10% larger than the latter.
在本实施例中,生成单元203可以将二值化运动图像中生成的感兴趣区域在视频图像帧中的对应区域作为该视频图像帧的感兴趣区域,由此,提取单元101能够从视频图像帧中提取出感兴趣区域。In the present embodiment, the generating unit 203 may use a corresponding region of the region of interest generated in the binarized moving image in the video image frame as the region of interest of the video image frame, whereby the extracting unit 101 can obtain the video image from the video image. The region of interest is extracted from the frame.
图3的框301示出了根据本申请的提取单元101从视频图像帧300中提取出的感兴趣区域的边界。 Block 301 of FIG. 3 illustrates the boundaries of the region of interest extracted from the video image frame 300 by the extraction unit 101 in accordance with the present application.
在本实施例中,检测单元102可以基于提取单元101所提取出的感兴趣区域,在视频图像帧中进行对象检测。In the present embodiment, the detecting unit 102 can perform object detection in the video image frame based on the region of interest extracted by the extracting unit 101.
图8是检测单元102的一个示意图,如图8所示,检测单元102可以包括判断单元801和对象检测单元802。FIG. 8 is a schematic diagram of the detecting unit 102. As shown in FIG. 8, the detecting unit 102 may include a determining unit 801 and an object detecting unit 802.
在本实施例中,判断单元801用于判断视频图像帧中的感兴趣区域的数量是否小于或等于第三阈值,并且感兴趣区域的面积是否小于或等于第四阈值;对象检测单元802根据判断单元801的判断结果,在视频图像帧的感兴趣区域中或视频图像帧的整个图像范围中进行对象检测。In this embodiment, the determining unit 801 is configured to determine whether the number of regions of interest in the video image frame is less than or equal to a third threshold, and whether the area of the region of interest is less than or equal to a fourth threshold; the object detecting unit 802 determines The result of the determination by unit 801 is object detection in the region of interest of the video image frame or the entire image range of the video image frame.
在本实施例中,如果判断单元801判断为该视频图像帧中的感兴趣区域的数量小于或等于第三阈值,并且,该视频图像帧中感兴趣区域的面积总和小于或等于第四阈值,那么,对象检测单元802在该视频图像帧的各感兴趣区域中进行对象检测,由此,能够进行快速的对象检测。In this embodiment, if the determining unit 801 determines that the number of regions of interest in the video image frame is less than or equal to a third threshold, and the sum of the areas of the region of interest in the video image frame is less than or equal to a fourth threshold, Then, the object detecting unit 802 performs object detection in each region of interest of the video image frame, whereby fast object detection can be performed.
此外,如果判断单元801判断为该视频图像帧中的感兴趣区域的数量为0,那么,该对象检测单元802不对该该视频图像帧进行对象检测。Furthermore, if the judging unit 801 determines that the number of regions of interest in the video image frame is 0, the object detecting unit 802 does not perform object detection on the video image frame.
在本实施例中,如果判断单元801判断为该视频图像帧中的感兴趣区域的数量大 于第三阈值,或者该视频图像帧中感兴趣区域的面积总和大于第四阈值,那么,说明视频图像帧中,对象检测单元802在该视频图像帧的整个图像范围中进行对象检测,由此,能够防止漏检。In this embodiment, if the determining unit 801 determines that the number of regions of interest in the video image frame is large At a third threshold, or the sum of the areas of the region of interest in the video image frame is greater than a fourth threshold, then in the video image frame, the object detecting unit 802 performs object detection in the entire image range of the video image frame, thereby Can prevent missed inspections.
在本实施例中,对象检测单元802进行对象检测的具体方法可以参考现有技术,本实施例不再进行说明。In this embodiment, the specific method for the object detection unit 802 to perform the object detection may refer to the prior art, and is not described in this embodiment.
此外,在本实施例中,可以将视频中特定的视频图像帧作为关键帧(key frame)而将视频中其它的视频图像帧作为普通帧(normal frame),其中,可以将间隔预定时间的视频图像帧或间隔预定数量帧的视频图像帧作为关键帧,此外,也可以采用其他方式来设定关键帧。判断单元801可以判断该视频图像帧是否为普通帧,对于普通帧,可以进一步根据判断单元801的判断结果来确定在该普通帧的感兴趣区域中或该普通帧的整个图像范围中进行对象检测;而对于关键帧,可以不用判断单元801进行进一步判断,而直接在该关键帧的整个图像范围中进行对象检测。由此,通过对关键帧进行整个图像范围中的对象检测,能够防止漏检。In addition, in this embodiment, a specific video image frame in the video may be used as a key frame, and other video image frames in the video may be used as a normal frame, wherein a video separated by a predetermined time may be used. The image frame or the video image frame of a predetermined number of frames is used as a key frame. In addition, other methods may be used to set the key frame. The determining unit 801 can determine whether the video image frame is a normal frame, and for the normal frame, further determining, according to the determination result of the determining unit 801, performing object detection in the region of interest of the normal frame or the entire image range of the normal frame. For the key frame, the determination unit 801 can be used for further determination, and the object detection can be performed directly in the entire image range of the key frame. Thereby, it is possible to prevent missed detection by performing object detection in the entire image range on the key frame.
在本实施例中,检测单元102还可以具有合并单元803。在对象检测单元802对当前视频图像帧的感兴趣区域进行对象检测的情况下,合并单元803可以将当前视频图像帧的感兴趣区域中的检测结果,以及针对当前视频图像帧之前的视频图像帧的检测结果进行合并,该合并后的检测结果例如可以包括:对当前视频图像帧的感兴趣区域的检测结果,以及当前视频图像帧的感兴趣区域之外的、对之前的视频图像帧的检测结果。In this embodiment, the detecting unit 102 may further have a merging unit 803. In the case where the object detecting unit 802 performs object detection on the region of interest of the current video image frame, the merging unit 803 may detect the detection result in the region of interest of the current video image frame and the video image frame before the current video image frame. The detection result is merged, and the combined detection result may include, for example, a detection result of the region of interest of the current video image frame, and detection of the previous video image frame outside the region of interest of the current video image frame. result.
图9是对检测结果进行合并的一个示意图,901是当前的视频图像帧902之前的视频图像帧,9011、9012是在视频图像帧901中检测出的对象物体,当前的视频图像帧902的感兴趣区域为9021,在感兴趣区域9021中检测到对象物体9022,当前的视频图像帧902的检测结果与之前的视频图像帧901的检测结果进行合并,得到合并的检测结果903,在该合并的检测结果903包括:当前的视频图像帧902中感兴趣区域9021中检测到的对象物体9022,以及当前的视频图像帧902的感兴趣区域9021之外的、在之前的视频图像帧901中检测出的对象物体9012。FIG. 9 is a schematic diagram of merging detection results, 901 is a video image frame before the current video image frame 902, and 9011, 9012 are object objects detected in the video image frame 901, and the current video image frame 902 is sensed. The region of interest is 9021, the target object 9022 is detected in the region of interest 9021, and the detection result of the current video image frame 902 is combined with the detection result of the previous video image frame 901 to obtain a combined detection result 903, in the merged The detection result 903 includes: the object object 9022 detected in the region of interest 9021 in the current video image frame 902, and the region other than the region of interest 9021 of the current video image frame 902, detected in the previous video image frame 901 Object object 9012.
下面,结合图10来说明检测单元102的工作流程。Next, the workflow of the detecting unit 102 will be described with reference to FIG.
步骤1001、判断单元801判断当前的视频图像帧是否为普通帧,如果是,则进行到步骤1002,如果否,则进行到步骤1005。 Step 1001: The determining unit 801 determines whether the current video image frame is a normal frame, and if yes, proceeds to step 1002, and if no, proceeds to step 1005.
步骤1002、判断单元801判断当前的视频图像帧中感兴趣区域的数量是否小于或等于第三阈值,如果是,则进行到步骤1003,如果否,则进行到步骤1005。Step 1002: The determining unit 801 determines whether the number of regions of interest in the current video image frame is less than or equal to a third threshold, and if yes, proceeds to step 1003, and if no, proceeds to step 1005.
步骤1003、判断单元801判断当前的视频图像帧中感兴趣区域的总面积是否小于或等于第四阈值,如果是,则进行到步骤1004,如果否,则进行到步骤1005。Step 1003: The determining unit 801 determines whether the total area of the region of interest in the current video image frame is less than or equal to the fourth threshold. If yes, proceed to step 1004. If no, proceed to step 1005.
步骤1004、对象检测单元802在该视频图像帧的感兴趣区域中进行对象检测。Step 1004: The object detecting unit 802 performs object detection in the region of interest of the video image frame.
步骤1005、对象检测单元802在该视频图像帧的整个图像范围中进行对象检测。Step 1005: The object detecting unit 802 performs object detection in the entire image range of the video image frame.
步骤1006、合并单元803将当前视频图像帧的感兴趣区域的检测结果与之前的视频图像帧的检测结果进行合并。Step 1006: The merging unit 803 combines the detection result of the region of interest of the current video image frame with the detection result of the previous video image frame.
根据本实施例,对象检测装置能够基于视频图像帧的运动信息来提取感兴趣区域,并根据提取出的感兴趣区域来进行对象检测,由此,能够针对每一个视频图像帧来更加准确地提取相应的感兴趣区域,从而提高对象检测的准确性,并提高检测速度。According to the present embodiment, the object detecting apparatus can extract the region of interest based on the motion information of the video image frame, and perform object detection based on the extracted region of interest, thereby being able to extract more accurately for each video image frame Corresponding regions of interest, thereby improving the accuracy of object detection and increasing the speed of detection.
实施例2Example 2
本申请实施例还提供一种对象检测方法,用于从视频图像帧中检测出对象物体,与实施例1的对象检测装置对应。The embodiment of the present application further provides an object detecting method for detecting a target object from a video image frame, corresponding to the object detecting device of Embodiment 1.
图11是本实施例2的对象检测方法的一个流程示意图,如图11所示,该检测方法可以包括:FIG. 11 is a schematic flowchart of the object detecting method in the second embodiment. As shown in FIG. 11, the detecting method may include:
步骤1101、基于视频图像帧的运动信息,从所述视频图像帧中提取出感兴趣区域;以及 Step 1101, extracting a region of interest from the video image frame based on motion information of a video image frame;
步骤1102、根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测。Step 1102: Perform object detection in the video image frame according to the extracted region of interest.
图12是本实施例2的提取出感兴趣区域的方法的一个示意图,如图12所示,该方法包括:FIG. 12 is a schematic diagram of a method for extracting a region of interest according to the second embodiment. As shown in FIG. 12, the method includes:
步骤1201、检测所述视频图像帧中的运动信息;Step 1201: Detect motion information in the video image frame.
步骤1202、根据所检测到的运动信息,划分出所述视频图像帧中的各运动物体所占据的区域;以及Step 1202: According to the detected motion information, divide an area occupied by each moving object in the video image frame;
步骤1203、根据所述视频图像帧中的各运动物体所占据的区域,生成至少一个感兴趣区域,所述至少一个感兴趣区域覆盖所述视频图像帧中的各运动物体所处的区域。Step 1203: Generate at least one region of interest according to an area occupied by each moving object in the video image frame, where the at least one region of interest covers an area where each moving object in the video image frame is located.
在本实施例的步骤1201中,可以基于前景检测来生成所述视频图像帧的二值化 运动图像(motion image),从而获得所述视频图像帧的所述运动信息。In step 1201 of this embodiment, binarization of the video image frame may be generated based on foreground detection. A motion image to obtain the motion information of the video image frame.
在本实施例的步骤1202中,可以对二值化运动图像进行连通域分割处理,以得到像素的至少一个连通域,所述至少一个连通域对应于所述视频图像帧中的各运动物体所占据的区域。In step 1202 of the embodiment, the connected domain segmentation process may be performed on the binarized moving image to obtain at least one connected domain of the pixel, where the at least one connected domain corresponds to each moving object in the video image frame. Occupied area.
在本实施例的步骤1202中,还可以生成各所述连通域的外接多边形。In step 1202 of this embodiment, an circumscribed polygon of each of the connected domains may also be generated.
在本实施例的步骤1202中,还可以将彼此距离小于或等于第一阈值的所述连通域合并为一个新的连通域。In step 1202 of this embodiment, the connected domains whose distances from each other are less than or equal to the first threshold may also be merged into one new connected domain.
在本实施例的步骤1203中,可以根据所述区域的距离,生成所述至少一个感兴趣区域。In step 1203 of the embodiment, the at least one region of interest may be generated according to the distance of the region.
图13是本实施例2的根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测的方法的一个示意图,如图13所示,该方法包括:FIG. 13 is a schematic diagram of a method for performing object detection in the video image frame according to the extracted region of interest according to the second embodiment. As shown in FIG. 13, the method includes:
步骤1301、判断所述视频图像帧中的所述感兴趣区域的数量是否小于或等于第三阈值,并且所述感兴趣区域的面积是否小于或等于第四阈值;以及Step 1301: Determine whether the number of the regions of interest in the video image frame is less than or equal to a third threshold, and whether an area of the region of interest is less than or equal to a fourth threshold;
步骤1302、根据所述判断的结果,在所述视频图像帧的所述感兴趣区域中或所述视频图像帧的整个图像范围中进行对象检测。Step 1302: Perform object detection in the region of interest of the video image frame or the entire image range of the video image frame according to the result of the determining.
如图13所示,该方法还包括:As shown in FIG. 13, the method further includes:
步骤1303、在对当前视频图像帧的感兴趣区域进行对象检测的情况下,将当前视频图像帧的感兴趣区域中的检测结果,以及针对当前视频图像帧之前的视频图像帧的检测结果进行合并。Step 1303: Combine the detection result in the region of interest of the current video image frame and the detection result of the video image frame before the current video image frame in the case of performing object detection on the region of interest of the current video image frame. .
关于上述个各骤的详细说明,可以参考实施例1中对于相应单元的说明,此处不再进行重复说明。For a detailed description of the above steps, reference may be made to the description of the corresponding unit in Embodiment 1, and the repeated description is not repeated here.
根据本实施例,对象检测方法能够基于视频图像帧的运动信息来提取感兴趣区域,并根据提取出的感兴趣区域来进行对象检测,由此,能够针对每一个视频图像帧来更加准确地提取相应的感兴趣区域,从而提高对象检测的准确性,并提高检测速度。According to the present embodiment, the object detecting method can extract the region of interest based on the motion information of the video image frame, and perform object detection based on the extracted region of interest, thereby being able to extract more accurately for each video image frame. Corresponding regions of interest, thereby improving the accuracy of object detection and increasing the speed of detection.
实施例3Example 3
本申请实施例3提供一种电子设备,包括如实施例1所述的对象检测装置。Embodiment 3 of the present application provides an electronic device including the object detecting device as described in Embodiment 1.
图14是本申请实施例3的电子设备的一个构成示意图。如图14所示,电子设备1400可以包括:中央处理器(CPU)1401和存储器1402;存储器1402耦合到中央 处理器1401。其中该存储器1402可存储各种数据;此外还存储进行对象检测的程序,并且在中央处理器1401的控制下执行该程序。FIG. 14 is a schematic diagram showing the configuration of an electronic device according to Embodiment 3 of the present application. As shown in FIG. 14, the electronic device 1400 can include a central processing unit (CPU) 1401 and a memory 1402; the memory 1402 is coupled to the center. The processor 1401. The memory 1402 can store various data; in addition, a program for performing object detection is stored, and the program is executed under the control of the central processing unit 1401.
在一个实施方式中,对象检测装置中的功能可以被集成到中央处理器1401中。In one embodiment, the functionality in the object detection device can be integrated into the central processor 1401.
其中,中央处理器1401可以被配置为:The central processing unit 1401 can be configured to:
基于视频图像帧的运动信息,从所述视频图像帧中提取出感兴趣区域;以及Extracting a region of interest from the video image frame based on motion information of the video image frame;
根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测。Object detection is performed in the video image frame based on the extracted region of interest.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
检测所述视频图像帧中的运动信息;Detecting motion information in the video image frame;
根据所检测到的运动信息,划分出所述视频图像帧中的各运动物体所占据的区域;以及Defining an area occupied by each moving object in the video image frame according to the detected motion information;
根据所述视频图像帧中的各运动物体所占据的区域,生成至少一个感兴趣区域,所述至少一个感兴趣区域覆盖所述视频图像帧中的各运动物体所处的区域。And generating at least one region of interest according to an area occupied by each moving object in the video image frame, the at least one region of interest covering an area in which each moving object in the video image frame is located.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
基于前景检测来生成所述视频图像帧的二值化运动图像(motion image),从而获得所述视频图像帧的所述运动信息。A binarized motion image of the video image frame is generated based on foreground detection, thereby obtaining the motion information of the video image frame.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
对所述二值化运动图像进行连通域分割处理,以得到像素的至少一个连通域,所述至少一个连通域对应于所述视频图像帧中的各运动物体所占据的区域。Connected domain segmentation processing is performed on the binarized moving image to obtain at least one connected domain of the pixel, the at least one connected domain corresponding to an area occupied by each moving object in the video image frame.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
生成各所述连通域的外接多边形。An circumscribed polygon of each of the connected domains is generated.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
将彼此距离小于或等于第一阈值的所述连通域合并为一个新的连通域。The connected domains that are less than or equal to the first threshold are merged into a new connected domain.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
根据所述区域的距离,生成所述至少一个感兴趣区域。The at least one region of interest is generated based on the distance of the region.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
判断所述视频图像帧中的所述感兴趣区域的数量是否小于或等于第三阈值,并且所述感兴趣区域的面积是否小于或等于第四阈值;以及Determining whether the number of the regions of interest in the video image frame is less than or equal to a third threshold, and whether an area of the region of interest is less than or equal to a fourth threshold;
根据所述判断的结果,在所述视频图像帧的所述感兴趣区域中或所述视频图像帧的整个图像范围中进行对象检测。 Based on the result of the determination, object detection is performed in the region of interest of the video image frame or in the entire image range of the video image frame.
中央处理器1401还可以被配置为:The central processor 1401 can also be configured to:
在对当前视频图像帧的感兴趣区域进行对象检测的情况下,将当前视频图像帧的感兴趣区域中的检测结果,以及针对当前视频图像帧之前的视频图像帧的检测结果进行合并。In the case of performing object detection on the region of interest of the current video image frame, the detection results in the region of interest of the current video image frame and the detection results of the video image frames preceding the current video image frame are combined.
此外,如图14所示,电子设备1400还可以包括:输入输出单元1403和显示单元1404等;其中,上述部件的功能与现有技术类似,此处不再赘述。值得注意的是,电子设备1400也并不是必须要包括图14中所示的所有部件;此外,电子设备1400还可以包括图14中没有示出的部件,可以参考现有技术。In addition, as shown in FIG. 14, the electronic device 1400 may further include: an input and output unit 1403, a display unit 1404, and the like; wherein the functions of the above components are similar to those of the prior art, and details are not described herein again. It should be noted that the electronic device 1400 does not necessarily have to include all the components shown in FIG. 14; in addition, the electronic device 1400 may further include components not shown in FIG. 14, and reference may be made to the prior art.
本申请实施例还提供一种计算机可读程序,其中当在对象检测装置或电子设备中执行所述程序时,所述程序使得所述对象检测装置或电子设备执行实施例2所述的对象检测方法。The embodiment of the present application further provides a computer readable program, wherein the program causes the object detecting device or the electronic device to perform the object detection described in Embodiment 2 when the program is executed in an object detecting device or an electronic device method.
本申请实施例还提供一种存储有计算机可读程序的存储介质,其中,所述存储介质存储上述计算机可读程序,所述计算机可读程序使得对象检测装置或电子设备执行实施例2所述的对象检测方法。The embodiment of the present application further provides a storage medium storing a computer readable program, wherein the storage medium stores the computer readable program, wherein the computer readable program causes the object detecting device or the electronic device to perform the embodiment 2 Object detection method.
结合本发明实施例描述的对象检测装置可直接体现为硬件、由处理器执行的软件模块或二者组合。例如,图1、2、8中所示的功能框图中的一个或多个和/或功能框图的一个或多个组合,既可以对应于计算机程序流程的各个软件模块,亦可以对应于各个硬件模块。这些软件模块,可以分别对应于实施例2所示的各个步骤。这些硬件模块例如可利用现场可编程门阵列(FPGA)将这些软件模块固化而实现。The object detecting apparatus described in connection with the embodiments of the present invention may be directly embodied as hardware, a software module executed by a processor, or a combination of both. For example, one or more of the functional blocks shown in Figures 1, 2, and 8 and/or one or more combinations of functional blocks may correspond to individual software modules of a computer program flow, or to individual hardware. Module. These software modules may correspond to the respective steps shown in Embodiment 2, respectively. These hardware modules can be implemented, for example, by curing these software modules using a Field Programmable Gate Array (FPGA).
软件模块可以位于RAM存储器、闪存、ROM存储器、EPROM存储器、EEPROM存储器、寄存器、硬盘、移动磁盘、CD-ROM或者本领域已知的任何其它形式的存储介质。可以将一种存储介质耦接至处理器,从而使处理器能够从该存储介质读取信息,且可向该存储介质写入信息;或者该存储介质可以是处理器的组成部分。处理器和存储介质可以位于ASIC中。该软件模块可以存储在移动终端的存储器中,也可以存储在可插入移动终端的存储卡中。例如,若设备(例如移动终端)采用的是较大容量的MEGA-SIM卡或者大容量的闪存装置,则该软件模块可存储在该MEGA-SIM卡或者大容量的闪存装置中。The software module can reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, removable disk, CD-ROM, or any other form of storage medium known in the art. A storage medium can be coupled to the processor to enable the processor to read information from, and write information to, the storage medium; or the storage medium can be an integral part of the processor. The processor and the storage medium can be located in an ASIC. The software module can be stored in the memory of the mobile terminal or in a memory card that can be inserted into the mobile terminal. For example, if a device (such as a mobile terminal) uses a larger capacity MEGA-SIM card or a large-capacity flash memory device, the software module can be stored in the MEGA-SIM card or a large-capacity flash memory device.
针对图1、2、8描述的功能框图中的一个或多个和/或功能框图的一个或多个组 合,可以实现为用于执行本申请所描述功能的通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现场可编程门阵列(FPGA)或其它可编程逻辑器件、分立门或晶体管逻辑器件、分立硬件组件、或者其任意适当组合。针对图1-3描述的功能框图中的一个或多个和/或功能框图的一个或多个组合,还可以实现为计算设备的组合,例如,DSP和微处理器的组合、多个微处理器、与DSP通信结合的一个或多个微处理器或者任何其它这种配置。One or more of the functional block diagrams described with respect to Figures 1, 2, 8 and/or one or more groups of functional block diagrams A general purpose processor, digital signal processor (DSP), application specific integrated circuit (ASIC), field programmable gate array (FPGA) or other programmable logic device, discrete door for performing the functions described herein can be implemented. Or transistor logic device, discrete hardware component, or any suitable combination thereof. One or more of the functional blocks described with respect to Figures 1-3 and/or one or more combinations of functional blocks may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, multiple microprocessors One or more microprocessors in conjunction with DSP communication or any other such configuration.
以上结合具体的实施方式对本申请进行了描述,但本领域技术人员应该清楚,这些描述都是示例性的,并不是对本申请保护范围的限制。本领域技术人员可以根据本申请的原理对本申请做出各种变型和修改,这些变型和修改也在本申请的范围内。 The present invention has been described in connection with the specific embodiments thereof, but it is to be understood that the description is intended to be illustrative and not restrictive. Various modifications and alterations of this application will be apparent to those skilled in the art in the light of the invention.

Claims (19)

  1. 一种对象检测(object detection)装置,用于从视频图像帧中检测出对象物体,该装置包括:An object detection device for detecting a target object from a video image frame, the device comprising:
    提取单元,其基于视频图像帧的运动信息,从所述视频图像帧中提取出感兴趣区域;以及An extracting unit that extracts a region of interest from the video image frame based on motion information of a video image frame;
    检测单元,其根据所述提取单元所提取出的感兴趣区域,在所述视频图像帧中进行对象检测。And a detecting unit that performs object detection in the video image frame according to the region of interest extracted by the extracting unit.
  2. 如权利要求1所述的对象检测装置,其中,所述提取单元包括:The object detecting device according to claim 1, wherein said extracting unit comprises:
    运动检测单元,其用于检测所述视频图像帧中的运动信息;a motion detecting unit, configured to detect motion information in the video image frame;
    区域划分单元,其用于根据运动检测单元所检测到的运动信息,划分出所述视频图像帧中的各运动物体所占据的区域;以及a region dividing unit configured to divide, according to motion information detected by the motion detecting unit, an area occupied by each moving object in the video image frame;
    生成单元,其根据所述视频图像帧中的各运动物体所占据的区域,生成至少一个感兴趣区域,所述至少一个感兴趣区域覆盖所述视频图像帧中的各运动物体所处的区域。And a generating unit that generates at least one region of interest according to an area occupied by each moving object in the video image frame, the at least one region of interest covering an area in which each moving object in the video image frame is located.
  3. 如权利要求2所述的对象检测装置,其中,The object detecting device according to claim 2, wherein
    所述运动检测单元基于前景检测来生成所述视频图像帧的二值化运动图像(motion image),从而获得所述视频图像帧的所述运动信息。The motion detecting unit generates a binarized motion image of the video image frame based on foreground detection, thereby obtaining the motion information of the video image frame.
  4. 如权利要求3所述的对象检测装置,其中,The object detecting device according to claim 3, wherein
    所述区域划分单元对所述二值化运动图像进行连通域分割处理,以得到像素的至少一个连通域,所述至少一个连通域对应于所述视频图像帧中的各运动物体所占据的区域。The area dividing unit performs a connected domain segmentation process on the binarized moving image to obtain at least one connected domain of the pixel, where the at least one connected domain corresponds to an area occupied by each moving object in the video image frame .
  5. 如权利要求4所述的对象检测装置,其中,The object detecting device according to claim 4, wherein
    所述区域划分单元生成各所述连通域的外接多边形。The area dividing unit generates a circumscribed polygon of each of the connected domains.
  6. 如权利要求4所述的对象检测装置,其中,The object detecting device according to claim 4, wherein
    所述区域划分单元将彼此距离小于或等于第一阈值的所述连通域合并为一个新的连通域。The area dividing unit merges the connected domains whose distances from each other by less than or equal to the first threshold into a new connected domain.
  7. 如权利要求2所述的对象检测装置,其中,The object detecting device according to claim 2, wherein
    所述生成单元根据所述区域的距离,生成所述至少一个感兴趣区域。 The generating unit generates the at least one region of interest according to the distance of the region.
  8. 如权利要求1所述的对象检测装置,其中,所述检测单元包括:The object detecting device according to claim 1, wherein said detecting unit comprises:
    判断单元,其用于判断所述视频图像帧中的所述感兴趣区域的数量是否小于或等于第三阈值,并且所述感兴趣区域的面积是否小于或等于第四阈值;以及a determining unit, configured to determine whether the number of the regions of interest in the video image frame is less than or equal to a third threshold, and whether an area of the region of interest is less than or equal to a fourth threshold;
    对象检测单元,其根据所述判断单元的判断结果,在所述视频图像帧的所述感兴趣区域中或所述视频图像帧的整个图像范围中进行对象检测。An object detecting unit that performs object detection in the region of interest of the video image frame or the entire image range of the video image frame according to a determination result of the determining unit.
  9. 如权利要求8所述的对象检测装置,其中,所述检测单元还包括:The object detecting device according to claim 8, wherein the detecting unit further comprises:
    合并单元,在所述对象检测单元对当前视频图像帧的感兴趣区域进行对象检测的情况下,将当前视频图像帧的感兴趣区域中的检测结果,以及针对当前视频图像帧之前的视频图像帧的检测结果进行合并。a merging unit, where the object detecting unit performs object detection on the region of interest of the current video image frame, the detection result in the region of interest of the current video image frame, and the video image frame before the current video image frame The test results are combined.
  10. 一种电子设备,其具有权利要求1-9中任一项所述的对象检测装置。An electronic device having the object detecting device according to any one of claims 1-9.
  11. 一种对象检测方法,用于从视频图像帧中检测出对象物体,该方法包括:An object detecting method for detecting a target object from a video image frame, the method comprising:
    基于视频图像帧的运动信息,从所述视频图像帧中提取出感兴趣区域;以及Extracting a region of interest from the video image frame based on motion information of the video image frame;
    根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测。Object detection is performed in the video image frame based on the extracted region of interest.
  12. 如权利要求11所述的对象检测方法,其中,从所述视频图像帧中提取出感兴趣区域包括:The object detecting method according to claim 11, wherein extracting the region of interest from the video image frame comprises:
    检测所述视频图像帧中的运动信息;Detecting motion information in the video image frame;
    根据所检测到的运动信息,划分出所述视频图像帧中的各运动物体所占据的区域;以及Defining an area occupied by each moving object in the video image frame according to the detected motion information;
    根据所述视频图像帧中的各运动物体所占据的区域,生成至少一个感兴趣区域,所述至少一个感兴趣区域覆盖所述视频图像帧中的各运动物体所处的区域。And generating at least one region of interest according to an area occupied by each moving object in the video image frame, the at least one region of interest covering an area in which each moving object in the video image frame is located.
  13. 如权利要求12所述的对象检测方法,其中,检测所述视频图像帧中的运动信息包括:The object detecting method according to claim 12, wherein detecting the motion information in the video image frame comprises:
    基于前景检测来生成所述视频图像帧的二值化运动图像(motion image),从而获得所述视频图像帧的所述运动信息。A binarized motion image of the video image frame is generated based on foreground detection, thereby obtaining the motion information of the video image frame.
  14. 如权利要求13所述的对象检测方法,其中,根据所检测到的运动信息,划分出所述视频图像帧中的各运动物体所占据的区域包括:The object detecting method according to claim 13, wherein the area occupied by each moving object in the video image frame is divided according to the detected motion information, including:
    对所述二值化运动图像进行连通域分割处理,以得到像素的至少一个连通域,所述至少一个连通域对应于所述视频图像帧中的各运动物体所占据的区域。Connected domain segmentation processing is performed on the binarized moving image to obtain at least one connected domain of the pixel, the at least one connected domain corresponding to an area occupied by each moving object in the video image frame.
  15. 如权利要求14所述的对象检测方法,其中,根据所检测到的运动信息,划 分出所述视频图像帧中的各运动物体所占据的区域还包括:The object detecting method according to claim 14, wherein, based on the detected motion information, The area occupied by each moving object in the video image frame is further included:
    生成各所述连通域的外接多边形。An circumscribed polygon of each of the connected domains is generated.
  16. 如权利要求14所述的对象检测方法,其中,根据所检测到的运动信息,划分出所述视频图像帧中的各运动物体所占据的区域还包括:The object detecting method according to claim 14, wherein the dividing the area occupied by each moving object in the video image frame according to the detected motion information further comprises:
    将彼此距离小于或等于第一阈值的所述连通域合并为一个新的连通域。The connected domains that are less than or equal to the first threshold are merged into a new connected domain.
  17. 如权利要求12所述的对象检测方法,其中,从所述视频图像帧中提取出感兴趣区域包括:The object detecting method according to claim 12, wherein extracting the region of interest from the video image frame comprises:
    根据所述区域的距离,生成所述至少一个感兴趣区域。The at least one region of interest is generated based on the distance of the region.
  18. 如权利要求11所述的对象检测方法,其中,根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测包括:The object detecting method according to claim 11, wherein the detecting the object in the video image frame according to the extracted region of interest comprises:
    判断所述视频图像帧中的所述感兴趣区域的数量是否小于或等于第三阈值,并且所述感兴趣区域的面积是否小于或等于第四阈值;以及Determining whether the number of the regions of interest in the video image frame is less than or equal to a third threshold, and whether an area of the region of interest is less than or equal to a fourth threshold;
    根据所述判断的结果,在所述视频图像帧的所述感兴趣区域中或所述视频图像帧的整个图像范围中进行对象检测。Based on the result of the determination, object detection is performed in the region of interest of the video image frame or in the entire image range of the video image frame.
  19. 如权利要求18所述的对象检测方法,其中,根据所提取出的感兴趣区域,在所述视频图像帧中进行对象检测还包括:The object detecting method according to claim 18, wherein performing object detection in the video image frame according to the extracted region of interest further comprises:
    在对当前视频图像帧的感兴趣区域进行对象检测的情况下,将当前视频图像帧的感兴趣区域中的检测结果,以及针对当前视频图像帧之前的视频图像帧的检测结果进行合并。 In the case of performing object detection on the region of interest of the current video image frame, the detection results in the region of interest of the current video image frame and the detection results of the video image frames preceding the current video image frame are combined.
PCT/CN2016/101204 2016-09-30 2016-09-30 Object detection method, object detection apparatus and electronic device WO2018058573A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2016/101204 WO2018058573A1 (en) 2016-09-30 2016-09-30 Object detection method, object detection apparatus and electronic device
CN201680087601.8A CN109479118A (en) 2016-09-30 2016-09-30 Method for checking object, object test equipment and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2016/101204 WO2018058573A1 (en) 2016-09-30 2016-09-30 Object detection method, object detection apparatus and electronic device

Publications (1)

Publication Number Publication Date
WO2018058573A1 true WO2018058573A1 (en) 2018-04-05

Family

ID=61762403

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/101204 WO2018058573A1 (en) 2016-09-30 2016-09-30 Object detection method, object detection apparatus and electronic device

Country Status (2)

Country Link
CN (1) CN109479118A (en)
WO (1) WO2018058573A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109584266A (en) * 2018-11-15 2019-04-05 腾讯科技(深圳)有限公司 A kind of object detection method and device
CN110738101A (en) * 2019-09-04 2020-01-31 平安科技(深圳)有限公司 Behavior recognition method and device and computer readable storage medium
CN111191730A (en) * 2020-01-02 2020-05-22 中国航空工业集团公司西安航空计算技术研究所 Method and system for detecting oversized image target facing embedded deep learning

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3819811A1 (en) * 2019-11-06 2021-05-12 Ningbo Geely Automobile Research & Development Co. Ltd. Vehicle object detection

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101198033A (en) * 2007-12-21 2008-06-11 北京中星微电子有限公司 Locating method and device for foreground image in binary image
CN101799968A (en) * 2010-01-13 2010-08-11 任芳 Detection method and device for oil well intrusion based on video image intelligent analysis
CN103020608A (en) * 2012-12-28 2013-04-03 南京荣飞科技有限公司 Method for identifying prisoner wears in prison video surveillance image
CN103971381A (en) * 2014-05-16 2014-08-06 江苏新瑞峰信息科技有限公司 Multi-target tracking system and method
CN104167004A (en) * 2013-05-16 2014-11-26 上海分维智能科技有限公司 Rapid moving vehicle detection method for embedded DSP platform
US20150131851A1 (en) * 2013-11-13 2015-05-14 Xerox Corporation System and method for using apparent size and orientation of an object to improve video-based tracking in regularized environments

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101325690A (en) * 2007-06-12 2008-12-17 上海正电科技发展有限公司 Method and system for detecting human flow analysis and crowd accumulation process of monitoring video flow
CN104573697B (en) * 2014-12-31 2017-10-31 西安丰树电子科技发展有限公司 Building hoist car demographic method based on Multi-information acquisition
CN105957110B (en) * 2016-06-29 2018-04-13 上海小蚁科技有限公司 Apparatus and method for detection object

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101198033A (en) * 2007-12-21 2008-06-11 北京中星微电子有限公司 Locating method and device for foreground image in binary image
CN101799968A (en) * 2010-01-13 2010-08-11 任芳 Detection method and device for oil well intrusion based on video image intelligent analysis
CN103020608A (en) * 2012-12-28 2013-04-03 南京荣飞科技有限公司 Method for identifying prisoner wears in prison video surveillance image
CN104167004A (en) * 2013-05-16 2014-11-26 上海分维智能科技有限公司 Rapid moving vehicle detection method for embedded DSP platform
US20150131851A1 (en) * 2013-11-13 2015-05-14 Xerox Corporation System and method for using apparent size and orientation of an object to improve video-based tracking in regularized environments
CN103971381A (en) * 2014-05-16 2014-08-06 江苏新瑞峰信息科技有限公司 Multi-target tracking system and method

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109584266A (en) * 2018-11-15 2019-04-05 腾讯科技(深圳)有限公司 A kind of object detection method and device
CN110738101A (en) * 2019-09-04 2020-01-31 平安科技(深圳)有限公司 Behavior recognition method and device and computer readable storage medium
CN110738101B (en) * 2019-09-04 2023-07-25 平安科技(深圳)有限公司 Behavior recognition method, behavior recognition device and computer-readable storage medium
CN111191730A (en) * 2020-01-02 2020-05-22 中国航空工业集团公司西安航空计算技术研究所 Method and system for detecting oversized image target facing embedded deep learning
CN111191730B (en) * 2020-01-02 2023-05-12 中国航空工业集团公司西安航空计算技术研究所 Method and system for detecting oversized image target oriented to embedded deep learning

Also Published As

Publication number Publication date
CN109479118A (en) 2019-03-15

Similar Documents

Publication Publication Date Title
CN109086691B (en) Three-dimensional face living body detection method, face authentication and identification method and device
US10192107B2 (en) Object detection method and object detection apparatus
CN110414507B (en) License plate recognition method and device, computer equipment and storage medium
JP6511149B2 (en) Method of calculating area of fingerprint overlap area, electronic device for performing the same, computer program, and recording medium
CN108875723B (en) Object detection method, device and system and storage medium
US9619708B2 (en) Method of detecting a main subject in an image
WO2021051604A1 (en) Method for identifying text region of osd, and device and storage medium
US9311533B2 (en) Device and method for detecting the presence of a logo in a picture
WO2018058595A1 (en) Target detection method and device, and computer system
US20190156499A1 (en) Detection of humans in images using depth information
TWI514327B (en) Method and system for object detection and tracking
TWI772757B (en) Object detection method, electronic device and computer-readable storage medium
WO2018058573A1 (en) Object detection method, object detection apparatus and electronic device
WO2019076187A1 (en) Video blocking region selection method and apparatus, electronic device, and system
JP2012038318A (en) Target detection method and device
CN109948521B (en) Image deviation rectifying method and device, equipment and storage medium
WO2018058530A1 (en) Target detection method and device, and image processing apparatus
TW201432620A (en) Image processor with edge selection functionality
JP6338429B2 (en) Subject detection apparatus, subject detection method, and program
CN111046845A (en) Living body detection method, device and system
US9947106B2 (en) Method and electronic device for object tracking in a light-field capture
CN108960247B (en) Image significance detection method and device and electronic equipment
CN109583266A (en) A kind of object detection method, device, computer equipment and storage medium
CN112183277A (en) Detection method and device for abandoned object and lost object, terminal equipment and storage medium
JP2016053763A (en) Image processor, image processing method and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16917323

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16917323

Country of ref document: EP

Kind code of ref document: A1