WO2022156763A1 - Procédé de détection d'objet cible et dispositif associé - Google Patents

Procédé de détection d'objet cible et dispositif associé Download PDF

Info

Publication number
WO2022156763A1
WO2022156763A1 PCT/CN2022/073151 CN2022073151W WO2022156763A1 WO 2022156763 A1 WO2022156763 A1 WO 2022156763A1 CN 2022073151 W CN2022073151 W CN 2022073151W WO 2022156763 A1 WO2022156763 A1 WO 2022156763A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
target object
target
evaluation
evaluation value
Prior art date
Application number
PCT/CN2022/073151
Other languages
English (en)
Chinese (zh)
Inventor
孔令广
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2022156763A1 publication Critical patent/WO2022156763A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection

Definitions

  • the present application relates to the field of monitoring, and in particular, to a target object detection method and device thereof.
  • a target object detection method and device thereof are proposed, which can reduce the data amount of the encoded image.
  • an embodiment of the present application provides a target object detection method, the method includes: acquiring a first image; detecting whether the first image includes a target object; When the target object is included, do not perform encoding on the first image; acquire a second image; detect whether the second image includes the target object; when the second image includes the target object, perform the encoding on the second image Encoding and sending the encoded second image to the monitoring platform server.
  • the cameras can perform relevant image processing on each image in the video stream after acquiring the video stream. Therefore, if the intelligent camera can identify valid images and transmit them to the monitoring platform server, not only It can reduce the data transmission pressure, and can reduce the storage pressure and data processing pressure of the monitoring platform server.
  • the target detection method performs target object detection on the images captured by the camera, and encodes only the images with the target objects, and the images without the target objects may not be encoded, thus reducing the number of encoded images.
  • images without a target object may not be sent to the monitoring platform server, thereby reducing the subsequent transmission amount and the storage cost of the monitoring platform server.
  • acquiring the first image includes: acquiring at least one image within a preset interval, where the preset interval includes a preset time interval and a preset a number interval; selecting a first image from the at least one image in an image selection manner, the method further comprising: determining not to perform encoding on the at least one image.
  • the processing method representing the image is taken as the processing method of the multiple images as a whole, and the processing method includes: encoding/non-encoding, sending to the monitoring platform server/not sending to the monitoring platform server.
  • the method further includes: using an evaluation method to determine an evaluation value of the second image, where the evaluation value is used to describe the quality of the second image. Image Quality.
  • an evaluation method may be used to determine the evaluation value of the second image that already has the target object. That is, evaluating the second image by the quantized value (evaluation value) enables the user to evaluate the first image more intuitively.
  • the evaluation method includes: when the second image includes multiple target objects, the evaluation value of the second image is the same as the first The sub-evaluation values of the multiple target objects in the two images are correlated.
  • the evaluation of the second image involves each target object in the second image.
  • this method takes into account the correlation between the multiple target objects, so that the value of the second image can be more accurately measured.
  • the evaluation value is related to one or more of the following: the definition of the target object area where the target object is located; the pixels of the target object area Quantity; the shooting angle of the target object area and the number of key features possessed by the target object.
  • the second image may be evaluated from one or more of the above four aspects, so that the second image can be more accurately evaluated.
  • the method further includes: judging whether the evaluation value satisfies a preset threshold; if it satisfies the preset threshold, storing the second image Sent to the monitoring platform server.
  • the method can individually send images with high image quality to the monitoring platform server, so that the monitoring platform server can perform separate/focus analysis on these images, thereby reducing the processing pressure of the monitoring platform server and improving processing efficiency.
  • the method further includes discarding the first image.
  • the first image can be discarded, thereby saving storage space.
  • embodiments of the present application provide a camera, where the camera includes: a lens for receiving light for generating an image; a camera body for implementing the first aspect or multiple possibilities of the first aspect One or several target object detection methods in the implementation manner.
  • the embodiments provide a target object detection device, the device includes: an image acquisition unit for acquiring at least one image; a target object detection unit for performing target object detection on the at least one image, A target image with a target object is determined, and images without a target object are discarded; an encoding unit is used to perform encoding on the target image to generate an encoded image; and a sending unit is used to send the encoded image to the monitoring platform server.
  • the device further includes: an image quality evaluation unit, configured to perform evaluation on the image quality of the target image, and determine an evaluation value of the target image .
  • the image quality evaluation unit is specifically configured to use the sub-evaluation values of the multiple target objects when the target image includes multiple target objects The evaluation value is determined.
  • the image quality evaluation unit is further configured to determine whether the evaluation value satisfies a preset threshold; The target image is sent to the cache unit.
  • the device further includes: a cache unit, configured to store a target image that meets the preset threshold.
  • the sending unit is further configured to send the image in the cache unit to the monitoring platform server.
  • the evaluation value is related to one or more of the following: the definition of the target object area where the target object is located; the pixels of the target object area number; the shooting angle of the target object area; and the number of key features possessed by the target object.
  • the target object detection unit is specifically configured to select a representative image from the at least one image; perform target object detection on the representative image; If the representative image has a target object, it is determined that the at least one image is a target image with the target object.
  • an embodiment of the present application provides a camera, including: a lens for collecting light; a sensor for generating an image by performing photoelectric conversion on the light collected by the lens; a processor and a processor cluster for executing the above The first aspect or one or more target object detection methods in multiple possible implementation manners of the first aspect.
  • embodiments of the present application provide a non-volatile computer-readable storage medium on which computer program instructions are stored, characterized in that, when the computer program instructions are executed by a processor, the above-mentioned first aspect is implemented Or one or more target object detection methods in multiple possible implementation manners of the first aspect.
  • FIG. 1 shows a schematic diagram of an application scenario according to an embodiment of the present application
  • FIG. 2 shows a diagram of data processing of a smart camera according to an embodiment of the present application
  • FIG. 3 shows a schematic structural diagram of a target detection system according to an embodiment of the present application
  • FIG. 4 shows a flow chart of steps of a target object detection method according to an embodiment of the present application
  • FIG. 5 shows a block diagram of a target object detection apparatus according to an embodiment of the present application.
  • “/” may indicate that the objects associated before and after are an “or” relationship, for example, A/B may indicate A or B; “and/or” may be used to describe that there are three types of associated objects A relationship, for example, A and/or B, can mean that A exists alone, A and B exist at the same time, and B exists alone, where A and B can be singular or plural.
  • words such as “first” and “second” may be used to distinguish technical features with the same or similar functions. The words “first”, “second” and the like do not limit the quantity and execution order, and the words “first”, “second” and the like do not limit the difference.
  • words such as “exemplary” or “for example” are used to represent examples, illustrations or illustrations, and any embodiment or design solution described as “exemplary” or “for example” should not be construed are preferred or advantageous over other embodiments or designs.
  • the use of words such as “exemplary” or “such as” is intended to present the relevant concepts in a specific manner to facilitate understanding.
  • the technical solution of the present application is applicable to the field of video surveillance, and video surveillance is an important part of a security protection system.
  • video surveillance is an important part of a security protection system.
  • the application scenario of the technical solution will be briefly described below with reference to FIG. 1 .
  • FIG. 1 is a schematic diagram of an application scenario to which the technical solution provided by the present application is applicable.
  • a video surveillance system can be used to monitor road conditions. Before performing video surveillance, it is necessary to set the target object of video surveillance.
  • target objects include pedestrians, non-motor vehicles, and motor vehicles.
  • the video surveillance system may include devices with audio/video capture functions and a surveillance platform server that performs data communication with these devices.
  • the video surveillance system includes only four video capture devices, but in practice, the video surveillance system may include more or less video/audio capture devices as needed.
  • the video capture device may be a camera
  • the camera may include a common camera and a smart camera
  • the common camera refers to a device that converts the captured video data into a suitable bit rate and uploads it to the monitoring platform server, that is, Ordinary cameras need to use the monitoring platform server to process the captured video data (data processing such as object recognition), while smart cameras can use the intelligent processing module embedded in them to first perform image processing on the video data, and then process the processed video data.
  • the video data is uploaded to the monitoring platform server, wherein the intelligent processing module may include modules such as a face recognition module, a license plate recognition module, and the like.
  • the camera of the present application includes a lens and a camera body.
  • the lens is used to receive the light used to generate the image.
  • the function of the lens is to present the light image of the observed target on the sensor of the camera, also known as optical imaging.
  • the lens combines various optical parts (reflectors, transmission mirrors, prisms) of different shapes and different media (plastic, glass or crystal) in a certain way, so that after the light is transmitted or reflected by these optical parts, according to people's expectations. It is necessary to change the transmission direction of the light to be received by the receiving device to complete the optical imaging process of the object.
  • each lens is composed of multiple groups of lenses with different curved curvatures combined at different intervals.
  • the focal length of the lens is determined by the selection of indicators such as spacing, lens curvature, and light transmittance.
  • the main parameters of the lens include: effective focal length, aperture, maximum image plane, field of view, distortion, relative illumination, etc. The value of each index determines the overall performance of the lens.
  • the camera body may include a sensor and a processor.
  • a sensor also known as an image sensor
  • CCD charge-coupled device
  • CMOS complementary metal oxide semiconductor
  • Both CCD and CMOS have a large number (eg, tens of millions) of photodiodes, each photodiode is called a photosensitive cell, and each photosensitive cell corresponds to a pixel.
  • the photodiode converts the light signal into an electrical signal containing brightness (or brightness and color) after receiving light, and the image is reconstructed accordingly.
  • Bayer array is a common image sensor technology that can be used in CCD and CMOS.
  • Bayer array uses Bayer color filter to make different pixels only sensitive to one of the three primary colors of red, blue and green. These pixels are interleaved and then interpolated by demosaicing to restore the original image.
  • Bayer arrays can be applied to CCD or CMOS, and sensors using Bayer arrays are also called Bayer sensors.
  • sensor technologies such as X3 (developed by Foveon).
  • X3 technology uses three layers of photosensitive elements, each layer records one of the color channels of RGB, so it can capture all colors on one pixel. Image sensor.
  • the processor (aka image processor), such as a system-on-chip (SoC), is used to convert the image produced by the sensor into a three-channel format (such as YUV), improve the image quality, and detect whether there is a target object in the image, and is also used for Encode the image.
  • SoC system-on-chip
  • the above-mentioned smart processing module may be included in the processor.
  • there may be only one processor eg, a multi-function integrated SoC
  • a cluster composed of multiple processors eg, multiple processors including an ISP and an encoder.
  • the smart camera 201 can transmit the data to the monitoring platform server by using the communication network.
  • the data can be transmitted to the storage unit 202 (such as a hard disk) of the monitoring platform server and store the data in the storage unit 202 (such as a hard disk). in unit 202.
  • the monitoring platform server refers to a device that can receive the data sent by the camera, perform related processing on the data, and store the data.
  • the monitoring platform server may be a single computing device or multiple computing devices, such as a server, server cluster, public cloud/private cloud.
  • the video monitoring system can preset the data sent to the monitoring platform server.
  • the data sent to the monitoring platform server may include a video stream collected by the smart camera, a target image determined by an intelligent processing module, and a region of interest (ROI) in the target image.
  • the region of interest is often the region where the target object is located, so it can be called the target object region.
  • the smart camera can also send the set content to the storage unit 202 through the server, for example, the captured video stream and the recognized face image.
  • the smart camera may send to the monitoring platform an encoded video stream that only includes a target image of the target object and a target image that meets a preset image quality.
  • Figure 3 shows a schematic structural diagram of the target detection system.
  • the target detection system 300 includes a plurality of cameras 301 to 305 and a monitoring platform server 310, wherein each camera in the plurality of cameras 301 to 305 may be an ordinary camera or a smart camera, In the case where the cameras 301 to 305 are smart cameras, the smart processing modules embedded in each smart camera may be the same or different.
  • the cameras 301 to 305 can transmit the acquired video data to the monitoring platform server, and the interface connecting the monitoring platform server 310 and the cameras 301 to 305 can be wired or wireless communication.
  • the wired mode may include the transmission control protocol/internet protocol (TCP/IP) communication technology, the user datagram protocol (UDP) technology or the standard universal serial bus ( universal serial bus, USB) port, COM interface and other similar standard ports.
  • the wireless communication method may include technologies such as WiFi, Bluetooth, ZigBee or ultra wideband (UWB). The corresponding connection method can be selected according to the actual application scenario and the hardware form of the camera.
  • FIG. 4 shows a flow chart of steps of a target object detection method according to an embodiment of the present application.
  • the target object detection method shown in FIG. 4 can be executed by a smart camera in a target detection system.
  • step S410 a first image is acquired, wherein the first image is an image acquired by the smart camera described above, or an image acquired by an ordinary camera, and the image is sent to the corresponding smart camera.
  • the method may acquire at least one image within a preset interval, and select the first image from the at least one image in an image selection manner. That is to say, considering that real-time detection is very challenging for the processing capability of the smart camera, it is possible to acquire multiple images within a preset interval, and then select an image from the multiple images as a representative image.
  • the preset interval mentioned here may be a time interval, such as multiple frames of images captured within five seconds, or may be a number interval, such as 10 frames of images captured continuously.
  • the method of selecting a representative image from among the multiple images, for example, the image selection method may be to select an intermediate image as the representative image, or, for example, to select the first frame of image as the representative image, which is not limited in this application.
  • step S420 it is detected whether a target object is included in the first image, wherein the target object is preset, and in implementation, the type of the target object may be a pedestrian, a motor vehicle and/or a non-motor vehicle.
  • the number of the target objects may be a single target object or multiple target objects. In the case of multiple target objects, as long as one target object is detected, it can be determined that the first image includes the target object. For example, if the image includes multiple non-motor vehicles, or if the same image includes pedestrians, motor vehicles and non-motor vehicles at the same time, if it is detected that the first image includes non-motor vehicles, the It is determined that the first image includes the target object.
  • the set target object corresponds to the intelligent processing module embedded in the smart camera, that is, the smart camera has an intelligent processing module for detecting the target object, so that the intelligent processing module can be used to determine whether the first image includes the target object.
  • the intelligent processing module can be implemented by an SoC.
  • the intelligent processing module may indicate an artificial intelligence (AI) module corresponding to the target, and may also be a machine learning module or a deep learning module, etc., wherein the AI module refers to loading a large amount of data into a computing device And choose a model to "fit" the data so that the computing device comes up with predictions/inferences.
  • Models used by computing devices include both simple equations (such as equations for a straight line) and very complex logical/mathematical systems, once the model to be used is selected and adjusted (that is, the model is improved by adjustment), Computing devices use the model to learn patterns in the data. Finally, the model can be used to perform processing on the input data.
  • the AI module may be a module with corresponding target detection capability, and the model used by it may be determined by actual users or technicians, for example, models corresponding to face recognition, pedestrian recognition, and license plate recognition.
  • step S430 when the target object is not included in the first image, encoding is not performed on the first image. That is, in a case where it is determined that the first image does not include the target object, encoding is not performed on the first image.
  • the first image may be deleted (discarded), thereby reducing the subsequent transmission amount and the storage cost of the monitoring platform server.
  • encoding is not performed on multiple images within the preset interval represented by the first image, and all are deleted.
  • Encoding is the encoding of three-channel images (such as YUV format images) for ease of transmission and viewing by users.
  • images such as YUV format images
  • Encoding is the encoding of three-channel images (such as YUV format images) for ease of transmission and viewing by users.
  • images such as YUV format images
  • an image encoded in JPG format or a video encoded in H.264/H.265.
  • the method may further perform step S440 to acquire a second image. If it is determined that the second image includes the target object after performing step S420 in the method, step S450 is performed to perform step S450 on the second image.
  • the encoding is performed and the encoded second image is sent to the monitoring platform server.
  • the second image may be encoded in H.264/H.265 format and transmitted to the monitoring platform server.
  • the second image with the target object may also be described as the target image.
  • steps S410 and S440 are executed in parallel. In other embodiments, the two may also be performed sequentially, that is, steps S410, S420 and S430 are performed first, and then steps S440, S420 and S450 are performed.
  • the method further includes performing an evaluation on the image quality of the second image if it is determined that the second image includes the target object.
  • One evaluation method is to use a single target object in the image to evaluate the image, for example, for the image including only pedestrian A, or for the image including pedestrian A and pedestrian B, use the evaluation for pedestrian A/pedestrian B. Evaluation to represent the evaluation of the entire image. This approach ignores the association between multiple target objects. As an example, pedestrian A, pedestrian B, and pedestrian C cross the road together. After a period of time after pedestrian A and pedestrian B walk together, pedestrian A, pedestrian B, and pedestrian C walk together. During this process, if only the images including pedestrian A or If the relevant area in the image (that is, the area including pedestrian A) is evaluated, the correlation between multiple pedestrians in the image will be ignored, thus underestimating the importance of the image.
  • the method may utilize an evaluation method to determine an evaluation value of the second image, wherein the evaluation value may be used to describe the image quality of the second image.
  • the evaluation value of the second image may be determined by using an evaluation method.
  • the method may first determine the number of target objects included in the second image, and then determine a specific evaluation manner according to the number of target objects.
  • the evaluation method may include a single target object evaluation method in which only a single target object is included in the second image or a multi-target object evaluation method in which at least two target objects are included in the second image.
  • the above-mentioned intelligent processing module can be used to detect target objects, and then determine the number of detected target objects.
  • a single-target object evaluation method may be adopted.
  • the target object can be evaluated from different dimensions according to the evaluation index of image quality, and the evaluation value of the second image can be determined.
  • the evaluation value is related to the evaluation index.
  • the evaluation index includes the clarity of the target object area where the target is exclusively located, the number of key features possessed by the target object, the number of pixels in the target object area, and the shooting angle of the photographed target object area.
  • the sharpness of the target object area refers to the sharpness of each detail shadow and its boundary on the image, and in implementation, the sharpness can be used to measure each detail in the image.
  • the AI module needs to use the feature values of the key features of the target object in the process of recognizing the target object. Therefore, the number of key features possessed by the target object determines the recognition rate of the AI module. The more pixels in the target object area, the larger the area occupied by the target object in the image, which is more conducive to various image processing. , or the profile of the target subject being photographed.
  • the second image may be evaluated by using one or more of the above-mentioned evaluation indicators to determine the evaluation value of the second image.
  • the method determines whether the evaluation value is higher than a predetermined threshold, if higher than the predetermined threshold, the image quality of the second image is high, and if it is lower than the predetermined threshold, the image quality of the second image is not high. If the evaluation value is higher than the predetermined threshold, the second image is sent to the buffer unit, and if the evaluation value is lower than the predetermined threshold, the second image may be discarded. As an example, the method may send the image stored in the buffer unit together with the video stream generated by encoding to the monitoring platform server.
  • different grades may be divided according to the evaluation value, each grade corresponds to a range of evaluation values, and if the evaluation value of the second image falls within the range of evaluation values, the second image corresponds to this grade.
  • the evaluation value may be divided into five grades, which may include excellent, good, moderate, pass, and fail.
  • the method may send an image with an evaluation value above a certain level to the cache unit, and if it is lower than a certain level, discard the image, as an example, the certain level may be a good level. Finally, the images stored in the cache unit are individually sent to the monitoring platform server.
  • the method may adopt a multi-target object evaluation method to determine the evaluation value of the second image, wherein the multi-object
  • the object evaluation method is a method in which the evaluation value of the second image is determined after each target object included in the second image is evaluated separately. That is to say, the multi-target evaluation method is related to the sub-evaluation values of the plurality of target objects in the second image.
  • the evaluation value of the second image may be determined by using the sub-evaluation value obtained for each target object and the corresponding sub-evaluation weight.
  • a corresponding sub-evaluation value may be calculated for each target object in the plurality of target objects, that is, the target object area corresponding to each target object may be evaluated from different dimensions by using the evaluation indicators described above.
  • the sub-evaluation value of For example, in the case where it is determined that the second image includes the first target object, the second target object and the third target object, the first target object area corresponding to the first target object in the second image and the The second target object area corresponding to the second target object and the third target object corresponding to the third target object.
  • the sub-evaluation values of the first target object, the second target object and the third target object are calculated respectively.
  • index values corresponding to each evaluation index can be calculated, and the sub-evaluation values of the target object can be finally calculated by using these evaluation values.
  • the evaluation methods for each target object may be the same or different. As an example, different evaluation methods may be determined according to the category of the target object. For example, the evaluation method for pedestrians is different from the evaluation method for motor vehicles.
  • S indicates the evaluation value of the second image
  • a i indicates the sub-evaluation value of the ith target object in the second image
  • b i indicates the sub-evaluation weight of the ith target object.
  • the sub-evaluation weight may be a weight preset by a user. It may also be a weight determined according to the characteristic value of the target object. For example, different sub-evaluation weights may be assigned according to the category of the target object, and different sub-evaluation weights may also be assigned according to the size of each target object area, and so on.
  • the evaluation value is higher than a predetermined threshold. If the evaluation value is higher than the predetermined threshold, the image quality of the second image is high, and if it is lower than the predetermined threshold, the image quality of the second image is not high. If the evaluation value is higher than the predetermined threshold, the second image is sent to the buffer unit, and if the evaluation value is lower than the predetermined threshold, the second image may be discarded.
  • each grade corresponds to a range of evaluation values, and if the evaluation value of the second image falls within the range of evaluation values, the second image corresponds to this grade.
  • It can be divided into five grades according to the evaluation value, and the five grades can include excellent, good, medium, pass and fail.
  • an image whose evaluation value is above a certain level may be sent to the buffer unit, and if it is lower than a certain level, the image may be discarded.
  • the certain level may be a good level.
  • the method may send an image that meets the preset requirements together with a video stream generated by encoding to the monitoring platform server, that is, send the image stored in the cache unit to the monitoring platform server together with the video stream.
  • the above-mentioned terminal and the like include corresponding hardware structures and/or software modules for executing each function.
  • the embodiments of the present application can be implemented in hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software driving hardware depends on the specific application and design constraints of the technical solution. Experts may use different methods for each specific application to implement the described functions, but such implementation should not be considered beyond the scope of the embodiments of the present application.
  • each functional module may be divided corresponding to each function, or two or at least two functions may be integrated into one processing module.
  • the above-mentioned integrated modules can be implemented in the form of hardware, and can also be implemented in the form of software function modules. It should be noted that the division of modules in the embodiments of the present application is schematic, and is only a logical function division, and other division methods may be used in actual implementation.
  • FIG. 5 shows a block diagram of a target object detection device according to an embodiment of the present application.
  • the target object detection device 500 includes an image acquisition unit 510, a target object detection unit 520, an encoding unit 530 and a sending unit 540, wherein the image acquisition unit 510 is used for acquiring at least one image; the target object detection unit 520 is used for Performing target object detection on the at least one image, determining a target image with a target object, and discarding images without the target object; the encoding unit 530 is configured to perform encoding on the target image to generate an encoded image; the sending unit 540 is used for Send the encoded image to the monitoring platform server.
  • the target object detection device 500 further includes an image quality evaluation unit, configured to perform evaluation on the image quality of the target image, and determine the evaluation value of the target image.
  • an image quality evaluation unit configured to perform evaluation on the image quality of the target image, and determine the evaluation value of the target image.
  • the image quality evaluation unit is specifically configured to use the sub-evaluation values of the multiple target objects to determine the evaluation value when the target image includes multiple target objects.
  • the image quality evaluation unit is further configured to determine whether the target image satisfies a preset threshold; if the preset threshold is met, send the target image to a cache unit.
  • the target object detection device 500 further includes: a cache unit, configured to store the target image satisfying the preset threshold.
  • the sending unit 540 is further configured to send the image in the cache unit to the monitoring platform server.
  • the evaluation value is related to one or more of the following: the clarity of the target object area where the target object is located; the number of pixels in the target object area; the shooting angle for photographing the target object area and the The number of key features the target object possesses.
  • the target object detection unit 510 is specifically configured to select a representative image from the at least one image; perform target object detection on the representative image; if there is a target object in the representative image, determine the at least one An image is a target image with a target object.
  • An embodiment of the present application provides a camera, including: a lens for collecting light; a sensor for generating an image by performing photoelectric conversion on the light collected by the lens; a processor and a processor cluster for executing the above-mentioned method.
  • Embodiments of the present application provide a non-volatile computer-readable storage medium on which computer program instructions are stored, and when the computer program instructions are executed by a processor, implement the above method.
  • Computer program instructions can be executed by a video camera or by a general purpose computer.
  • Embodiments of the present application provide a computer program product, including computer-readable codes, or a non-volatile computer-readable storage medium carrying computer-readable codes, when the computer-readable codes are stored in a processor of an electronic device When running in the electronic device, the processor in the electronic device executes the above method.
  • a computer-readable storage medium may be a tangible device that can hold and store instructions for use by the instruction execution device.
  • the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
  • Computer-readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read-only memory (Electrically Programmable Read-Only-Memory, EPROM or flash memory), static random access memory (Static Random-Access Memory, SRAM), portable compact disk read-only memory (Compact Disc Read-Only Memory, CD - ROM), Digital Video Disc (DVD), memory sticks, floppy disks, mechanically encoded devices, such as punch cards or raised structures in grooves on which instructions are stored, and any suitable combination of the foregoing .
  • RAM random access memory
  • ROM read only memory
  • EPROM erasable programmable read-only memory
  • EPROM Errically Programmable Read-Only-Memory
  • SRAM static random access memory
  • portable compact disk read-only memory Compact Disc Read-Only Memory
  • CD - ROM Compact Disc Read-Only Memory
  • DVD Digital Video Disc
  • memory sticks floppy disks
  • the computer readable program instructions or code described herein may be downloaded to various computing/processing devices from a computer readable storage medium, or to an external computer or external storage device over a network such as the Internet, a local area network, a wide area network and/or a wireless network.
  • the network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
  • a network adapter card or network interface in each computing/processing device receives computer-readable program instructions from a network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in each computing/processing device .
  • the computer program instructions used to perform the operations of the present application may be assembly instructions, Instruction Set Architecture (ISA) instructions, machine instructions, machine-related instructions, microcode, firmware instructions, state setting data, or in one or more source or object code written in any combination of programming languages, including object-oriented programming languages such as Smalltalk, C++, etc., and conventional procedural programming languages such as the "C" language or similar programming languages.
  • the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server implement.
  • the remote computer can be connected to the user's computer through any kind of network—including a Local Area Network (LAN) or a Wide Area Network (WAN)—or, can be connected to an external computer (e.g. use an internet service provider to connect via the internet).
  • electronic circuits such as programmable logic circuits, Field-Programmable Gate Arrays (FPGA), or Programmable Logic Arrays (Programmable Logic Arrays), are personalized by utilizing state information of computer-readable program instructions.
  • Logic Array, PLA the electronic circuit can execute computer readable program instructions to implement various aspects of the present application.
  • These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processor of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
  • These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
  • Computer readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more functions for implementing the specified logical function(s) executable instructions.
  • the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in hardware (eg, circuits or ASICs (Application) that perform the corresponding functions or actions. Specific Integrated Circuit, application-specific integrated circuit)), or can be implemented by a combination of hardware and software, such as firmware.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Studio Devices (AREA)
  • Train Traffic Observation, Control, And Security (AREA)

Abstract

Procédé de détection d'objet cible et dispositif associé. Le procédé comprend la réalisation d'une détection d'objet cible sur des images photographiées par une caméra, une image ayant un objet cible étant codée et transmise à un serveur de plateforme de surveillance, et une image n'ayant pas d'objet cible n'étant pas codée et n'étant pas transmise au serveur de plateforme de surveillance.
PCT/CN2022/073151 2021-01-25 2022-01-21 Procédé de détection d'objet cible et dispositif associé WO2022156763A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110099193.0 2021-01-25
CN202110099193.0A CN114898239A (zh) 2021-01-25 2021-01-25 目标对象检测方法及其设备

Publications (1)

Publication Number Publication Date
WO2022156763A1 true WO2022156763A1 (fr) 2022-07-28

Family

ID=82548515

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/073151 WO2022156763A1 (fr) 2021-01-25 2022-01-21 Procédé de détection d'objet cible et dispositif associé

Country Status (2)

Country Link
CN (1) CN114898239A (fr)
WO (1) WO2022156763A1 (fr)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060078217A1 (en) * 2004-05-20 2006-04-13 Seiko Epson Corporation Out-of-focus detection method and imaging device control method
CN107291810A (zh) * 2017-05-18 2017-10-24 深圳云天励飞技术有限公司 数据处理方法、装置及存储介质
CN110868600A (zh) * 2019-11-11 2020-03-06 腾讯云计算(北京)有限责任公司 目标跟踪视频推流方法、显示方法、装置和存储介质
CN111340140A (zh) * 2020-03-30 2020-06-26 北京金山云网络技术有限公司 图像数据集的获取方法、装置、电子设备及存储介质

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060078217A1 (en) * 2004-05-20 2006-04-13 Seiko Epson Corporation Out-of-focus detection method and imaging device control method
CN107291810A (zh) * 2017-05-18 2017-10-24 深圳云天励飞技术有限公司 数据处理方法、装置及存储介质
CN110868600A (zh) * 2019-11-11 2020-03-06 腾讯云计算(北京)有限责任公司 目标跟踪视频推流方法、显示方法、装置和存储介质
CN111340140A (zh) * 2020-03-30 2020-06-26 北京金山云网络技术有限公司 图像数据集的获取方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN114898239A (zh) 2022-08-12

Similar Documents

Publication Publication Date Title
US10986338B2 (en) Thermal-image based video compression systems and methods
JP5882455B2 (ja) 高解像度マルチスペクトル画像キャプチャ
JP6732902B2 (ja) 撮像装置および撮像システム
TWI713794B (zh) 用於識別在動作視訊中之事件的方法
EP3579145A1 (fr) Procédé et dispositif de traitement d'images, support d'enregistrement lisible sur ordinateur et dispositif électronique
WO2019196539A1 (fr) Procédé et appareil de fusion d'image
US8798369B2 (en) Apparatus and method for estimating the number of objects included in an image
WO2020094088A1 (fr) Procédé de capture d'image, caméra de surveillance et système de surveillance
TWI522967B (zh) 基於小腦模型網路的移動物體偵測方法及其裝置
JP2016129347A (ja) コンテクストデータを使用して端末によって画像キャプチャの確率を自動的に決定する方法
CN107704798B (zh) 图像虚化方法、装置、计算机可读存储介质和计算机设备
CN103905727A (zh) 被摄体区域跟踪设备及其控制方法
WO2022237591A1 (fr) Procédé et appareil d'identification d'objet mobile, dispositif électronique et support de stockage lisible
JP7024736B2 (ja) 画像処理装置、および画像処理方法、並びにプログラム
CN107613216A (zh) 对焦方法、装置、计算机可读存储介质和电子设备
US20230127009A1 (en) Joint objects image signal processing in temporal domain
TWI521473B (zh) 影像分析裝置、方法及電腦可讀取媒體
US20120033854A1 (en) Image processing apparatus
CN110929615B (zh) 图像处理方法、图像处理装置、存储介质与终端设备
US9189863B2 (en) Method and system for detecting motion capable of removing shadow by heat
WO2022156763A1 (fr) Procédé de détection d'objet cible et dispositif associé
CN107959840A (zh) 图像处理方法、装置、计算机可读存储介质和计算机设备
CN111800605A (zh) 基于枪球联动的车形、车牌传输的方法及系统、设备
Law et al. Performance enhancement of PRNU-based source identification for smart video surveillance
US12035033B2 (en) DNN assisted object detection and image optimization

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22742242

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22742242

Country of ref document: EP

Kind code of ref document: A1